Refine
Has Fulltext
- yes (108) (remove)
Year of publication
- 2015 (108) (remove)
Document Type
- Doctoral Thesis (108) (remove)
Is part of the Bibliography
- yes (108)
Keywords
- Klimawandel (3)
- climate change (3)
- Arbeitsgedächtnis (2)
- Aspekt (2)
- Bakteriophagen (2)
- Erosion (2)
- Germany (2)
- Geschäftsprozessmanagement (2)
- Modellierung (2)
- Nanopartikel (2)
Institute
- Institut für Geowissenschaften (23)
- Institut für Physik und Astronomie (15)
- Institut für Biochemie und Biologie (11)
- Institut für Ernährungswissenschaft (7)
- Sozialwissenschaften (7)
- Institut für Chemie (6)
- Institut für Umweltwissenschaften und Geographie (6)
- Department Linguistik (4)
- Institut für Mathematik (4)
- Department Erziehungswissenschaft (3)
A main limitation in the field of flood hydrology is the short time period covered by instrumental flood time series, rarely exceeding more than 50 to 100 years. However, climate variability acts on short to millennial time scales and identifying causal linkages to extreme hydrological events requires longer datasets. To extend instrumental flood time series back in time, natural geoarchives are increasingly explored as flood recorders. Therefore, annually laminated (varved) lake sediments seem to be the most suitable archives since (i) lake basins act as natural sediment traps in the landscape continuously recording land surface processes including floods and (ii) individual flood events are preserved as detrital layers intercalated in the varved sediment sequence and can be dated with seasonal precision by varve counting.
The main goal of this thesis is to improve the understanding about hydrological and sedimentological processes leading to the formation of detrital flood layers and therewith to contribute to an improved interpretation of lake sediments as natural flood archives. This goal was achieved in two ways: first, by comparing detrital layers in sediments of two dissimilar peri-Alpine lakes, Lago Maggiore in Northern Italy and Mondsee in Upper Austria, with local instrumental flood data and, second, by tracking detrital layer formation during floods by a combined hydro-sedimentary monitoring network at Lake Mondsee spanning from the rain fall to the deposition of detrital sediment at the lake floor.
Successions of sub-millimetre to 17 mm thick detrital layers were detected in sub-recent lake sediments of the Pallanza Basin in the western part of Lago Maggiore (23 detrital layers) and Lake Mondsee (23 detrital layers) by combining microfacies and high-resolution micro X-ray fluorescence scanning techniques (µ-XRF). The detrital layer records were dated by detailed intra-basin correlation to a previously dated core sequence in Lago Maggiore and varve counting in Mondsee. The intra-basin correlation of detrital layers between five sediment cores in Lago Maggiore and 13 sediment cores in Mondsee allowed distinguishing river runoff events from local erosion. Moreover, characteristic spatial distribution patterns of detrital flood layers revealed different depositional processes in the two dissimilar lakes, underflows in Lago Maggiore as well as under- and interflows in Mondsee. Comparisons with runoff data of the main tributary streams, the Toce River at Lago Maggiore and the Griesler Ache at Mondsee, revealed empirical runoff thresholds above which the deposition of a detrital layer becomes likely. Whereas this threshold is the same for the whole Pallanza Basin in Lago Maggiore (600 m3s-1 daily runoff), it varies within Lake Mondsee. At proximal locations close to the river inflow detrital layer deposition requires floods exceeding a daily runoff of 40 m3s-1, whereas at a location 2 km more distal an hourly runoff of 80 m3s-1 and at least 2 days with runoff above 40 m3s-1 are necessary. A relation between the thickness of individual deposits and runoff amplitude of the triggering events is apparent for both lakes but is obviously further influenced by variable influx and lake internal distribution of detrital sediment.
To investigate processes of flood layer formation in lake sediments, hydro-sedimentary dynamics in Lake Mondsee and its main tributary stream, Griesler Ache, were monitored from January 2011 to December 2013. Precipitation, discharge and turbidity were recorded continuously at the rivers outlet to the lake and compared to sediment fluxes trapped close to the lake bottom on a basis of three to twelve days and on a monthly basis in three different water depths at two locations in the lake basin, in a distance of 0.9 (proximal) and 2.8 km (distal) to the Griesler Ache inflow. Within the three-year observation period, 26 river floods of different amplitude (10-110 m3s-1) were recorded resulting in variable sediment fluxes to the lake (4-760 g m-2d-1). Vertical and lateral variations in flood-related sedimentation during the largest floods indicate that interflows are the main processes of lake internal sediment transport in Lake Mondsee. The comparison of hydrological and sedimentological data revealed (i) a rapid sedimentation within three days after the peak runoff in the proximal and within six to ten days in the distal lake basin, (ii) empirical runoff thresholds for triggering sediment flux at the lake floor increasing from the proximal (20 m3s-1) to the distal lake basin (30 m3s-1) and (iii) factors controlling the amount of detrital sediment deposition at a certain location in the lake basin. The total influx of detrital sediment is mainly driven by runoff amplitude, catchment sediment availability and episodic sediment input by local sediment sources. A further role plays the lake internal sediment distribution which is not the same for each event but is favoured by flood duration and the existence of a thermocline and, therewith, the season in which a flood occurred.
In summary, the studies reveal a high sensitivity of lake sediments to flood events of different intensity. Certain runoff amplitudes are required to supply enough detrital material to form a visible detrital layer at the lake floor. Reasonable are positive feedback mechanisms between rainfall, runoff, erosion, fluvial sediment transport capacity and lake internal sediment distribution. Therefore, runoff thresholds for detrital layer formation are site-specific due to different lake-catchment characteristics. However, the studies also reveal that flood amplitude is not the only control for the amount of deposited sediment at a certain location in the lake basin even for the strongest flood events. The sediment deposition is rather influenced by a complex interaction of catchment and in-lake processes. This means that the coring location within a lake basin strongly determines the significance of a flood layer record. Moreover, the results show that while lake sediments provide ideal archives for reconstructing flood frequencies, the reconstruction of flood amplitudes is a more complex issue and requires detailed knowledge about relevant catchment and in-lake sediment transport and depositional processes.
Diese Arbeit befasst sich mit der ganzheitlichen Betrachtung der Fluideigenschaften eines unterpermischen Reservoirs am Geothermie Forschungsstandort Groß Schönebeck (GrSk) bei Reservoirbedingungen und im Betrieb der Geothermieanlage. Die Untersuchungen zur Fluidherkunft ergeben, dass es sich um ein konnates Wasser meteorischen Ursprungs ohne den Einfluss der darüberliegenden Zechsteinwässer handelt. Die Ionen und Isotopenverhältnisse im Formationswasser gelöster Komponenten in GrSk belegen einen gemeinsamen Genesepfad mit Wässern anderer Rotliegend-Reservoire des Nordostdeutschen Beckens (NEGB). Die Isotopenverhältnisse von ⁸⁷Sr/⁸⁶Sr ≈ 0,7158 und von δ³⁴SV CDT ≈ 4,1 ‰ des Sulfats weisen auf die Anreicherung des Fluids mit schweren Isotopen durch die Fluid Gestein-Wechselwirkung mit Vulkaniten und Rotliegend Sandsteinen des Unteren Perms hin.
Das im Formationswasser bei Reservoirbedingungen gelöste Gas (Gas/Wasser ≤ 2 bei STP) enthält Stickstoff (δ¹⁵NAir ≈ 0,6 ‰) und thermogenes Methan (δ¹³CV-PDB ≈ - 18 ‰) aus organischen Karbonablagerungen (Kerogen Typ - III Kohlen) hoher Reife. Die Isotopenverhältnisse der Edelgase belegen eine krustale Herkunft des Gasgemisches. Die berechnete Verweilzeit τ (⁴He) der Gase im Reservoir liegt zwischen 275 und 317 Ma und überschreitet damit bei gegebener Konzentration von Mutternukliden im Reservoirgestein das allgemein angenommene Zeitalter der Sedimentgruppe. Das lässt sich durch eine Zuwanderung von Gasen aus älteren Sedimentfolgen erklären.
Die Veränderungen der physikochemischen Fluidparameter während des Anlagenbetriebs sind hauptsächlich temperaturbedingt. Bei stabilen Produktionsbedingungen und einer Temperatur von ca. 100 °C stabilisieren sich auch die Fluideigenschaften. Bei In situ Bedingungen übertage beträgt die Dichte ρ = 1,1325 ± 0,0002 g ∙ mL⁻¹, das Redoxpotential Eh = -105,5 ± 1,3 mV und der pH = 6,61 ± 0,002. Die relative Zusammensetzung der Gasphase bei stabilen Produktionsbedingungen zeigt dagegen eine geringe Erhöhung des Stickstoffanteils sowie des Anteils der Kohlenwasserstoffe (Ethan, Propan, usw.) und Abnahme des relativen Methananteils im Laufe des Betriebs.
Die quantitative Untersuchung der sekundären mineralischen Ausfällungen im Fluid mittels sequentieller Extraktion zeigte, dass Schwermetalle als eine Hauptkomponente der Fluidfestphase größtenteils in Verbindung mit organischen Molekülen vorliegen. Experimente zum Einfluss organischer Verbindungen unterschiedlicher Substanzklassen auf eine Mobilisierung der Schwermetalle aus dem Reservoirgestein ergaben, dass die Verbindungen wie Fettsäuren und PAK (polyzyklische aromatische Kohlenwasserstoffe) die Freisetzung von Kupfer, Nickel, Chrom und Blei verhindern bzw. zu derer Immobilisierung beitragen. Im Gegensatz dazu wird die Mobilität von Zink in Anwesenheit von diesen Verbindungen erhöht. Niedermolekulare Monocarbonsäuren und stickstoffhaltige Heteroaromaten tragen, mit Ausnahme von Blei, zur Freisetzung bzw. Mobilisierung von Schwermetallen aus dem Reservoirgestein bei.
Die gewonnenen Erkenntnisse dieser Arbeit bestätigen das Risiko massiver Ausfällungen auf der kalten Seite der Geothermieanlage bei Inbetriebnahme des Kraftwerks, wenn keine an den Fluidchemismus angepassten Präventionsmethoden eingesetzt werden. Die Isotopenzusammensetzung der Fluidkomponenten sowie geringfügige Schwankungen der Gaszusammensetzung im kontinuierlichen Anlagenbetrieb lässt eine Kommunikation des unterpermischen Reservoirs mit dem darunter liegenden Oberkarbon vermuten, was eine nachträgliche Veränderung der Fluidzusammensetzung beim Dauerbetrieb der Anlage bedeuten kann.
Peak oil is forcing our society to shift from fossil to renewable resources. However, such renewable resources are also scarce, and they too must be used in the most efficient and sustainable way possible. Biorefining is a concept that represents both resource efficiency and sustainability. This approach initiates a cascade use, which means food and feed production before material use, and an energy-related use at the end of the value-added chain. However, sustainability should already start in the fields, on the agricultural side, where the industrially-used biomass is produced. Therefore, the aim of my doctoral thesis is to analyse the sustainable feedstock supply for biorefineries. In contrast to most studies on biorefineries, I focus on the sustainable provision of feedstock and not on the bioengineering processing of whatever feedstock is available.
Grasslands provide a high biomass potential. They are often inefficiently used, so a new utilisation concept based on the biorefining approach can increase the added value from grasslands. Fodder legumes from temporary and permanent grasslands were chosen for this study. Previous research shows that they are a promising feedstock for industrial uses, and their positive environmental impact is an important byproduct to promote sustainable agricultural production systems.
Green Biorefineries are a class of biorefineries that use fresh green biomass, such as grasses or fodder legumes, as feedstock. After fractionation, an organic solution (press juice) forms; this is used for the production of organic acids, chemicals and extracts, as well as fertilisers. A fibre component (press cake) is also created to produce feed, biomaterials and biogas. This thesis examines a specific value chain, using alfalfa and clover/grass as feedstock and generating lactic acid and one type of cattle feed from it. The research question is if biomass production needs to be adapted for the utilisation of fodder legumes in the Green Biorefinery approach. I have attempted to give a holistic analysis of cultivation, processing and utilisation of two specific grassland crops. Field trials with alfalfa and clover/grass at different study sites were carried out to obtain information on biomass quality and quantity depending on the crop, study site and harvest time. The fresh biomass was fractionated with a screw press and the composition of press juices and cakes was analysed. Fermentation experiments took place to determine the usability of press juices for lactic acid production. The harvest time is not of high importance for the quality of press juices as a fermentation medium. For permanent grasslands, late cuts, often needed for reasons of nature conservation, are possible without a major influence on feedstock quality. The press cakes were silaged for feed-value determination.
Following evidence that both intermediate products are suitable feedstocks in the Green Biorefinery approach, I developed a cost-benefit analysis, comparing different production scenarios on a farm. Two standard crop rotations for Brandenburg, producing either only market crops or market crops and fodder legumes for ruminant feed production, were compared to a system that uses the cultivated fodder legumes for the Green Biorefinery value chain instead of only feed production. Timely processing of the raw material is important to maintain quality for industrial uses, so on-site processing at the farm is assumed in Green Biorefinery scenario. As a result, more added value stays in the rural area. Two farm sizes, common for many European regions, were chosen to examine the influence of scale. The cost site of farmers has also been analysed in detail to assess which farm characteristics make production of press juices for biochemical industries viable. Results show that for large farm sizes in particular, the potential profits are high. Additionally, the wider spectrum of marketable products generates new sources of income for farmers.
The holistic analysis of the supply chain provides evidence that the cultivation processes for fodder legumes do not need to be adapted for use in Green Biorefineries. In fact, the new utilisation approach even widens the cultivation and processing spectrum and can increase economic viability of fodder legume production in conventional farming.
In der vorliegenden Dissertation werden Migrationsdiskurse in der deutschen Grenzregion zu Polen im Vorfeld der EU-Erweiterung zum 1. Mai 2004 exemplarisch in drei deutschen Grenzstädten, der jeweils westliche Teil der ehemals gemeinsamen, seit 1945 durch eine nationalstaatliche Grenze mit dem jeweils zeitspezifischen Grenzregime geteilten, deutsch-polnischen Zwillingsstädte Frankfurt (Oder) – Słubice, Guben – Gubin und Görlitz – Zgorzelec, analysiert.
Ausgewählt wurde der Untersuchungsraum mit Blick auf die tiefgreifenden europäischen Transformationsprozesse seit den späten 1980er Jahren, die für die örtliche Bevölkerung gravierende lebensweltliche Strukturumbrüche zur Folge hatten. Die Region wurde mit der Vereinigung der beiden deutschen Staaten überdies zu einem zentralen Aktionsraum nationaler und internationaler Migrationspolitik; ihr wurde eine wichtige stellvertretende Funktion betreffend die Zutrittsregelung zugewiesen. Mit der EU-Erweiterung waren für die Region neuerliche, unmittelbare Veränderungen verbunden, die vor Ort gerade auch aufgrund damit (mutmaßlich) einhergehender Migration eher als Bedrohung denn als Chance gedeutet wurden.
Den diskurstheoretischen Hintergrund der Untersuchungen stellen in erster Linie die Arbeiten von Michel Foucault und die von Siegfried Jäger darauf aufruhend konzipierte Kritische Diskursanalyse bereit. Diskurs wird – grob vereinfacht – als Fluss von sozialen Wissensbeständen und Bewusstseinsinhalten durch die Zeit verstanden, der individuelles und kollektives Handeln von Menschen bestimmt; Diskurse sind der Ort, an dem (Be-)Deutungen von Menschen ausgehandelt, verändert und der Wirklichkeit zugewiesen werden. Der Forschungszugang versteht sich als Teil der Neuen Kulturgeographie, die konsequent nicht-essentialistisch und erkenntnistheoretisch nicht-fundamentalistisch ist.
Die Datenbasis der empirischen Analysen repräsentieren zwei Ebenen bzw. Teilsektoren des Diskurses. Zum einen die Berichterstattung der jeweils monopolartigen regionalen Tageszeitung in Frankfurt (Oder), Guben und Görlitz (Märkische Oderzeitung/Frankfurter Stadtbote, Lausitzer Rundschau/Lokalausgabe Guben, Sächsische Zeitung/Görlitzer Zeitung). Zum anderen ein Sample von insgesamt 17 Experteninterviews mit lokalen Funktionsträgern, die mit Blick auf ihr, an ihre spezifische professionelle und/oder ehrenamtliche Tätigkeit gebundenes, praxisgesättigtes Sonder- bzw. Insiderwissen zum Thema Migration ausgewählt und befragt wurden.
Die durchgeführten Analysen verdeutlichen unter anderem die Bedeutung diskurssemantischer Grundfiguren des deutschen Migrationsdiskurses im Sinne politisch und alltagskulturell konservierter migrationskritischer Vorstellungsinhalte und Bedeutungszuweisungen zu Kategorien des Fremden und Konstruktionen von Wir und/vs. Sie. Ebenso explizieren sie eine gravierende Diskrepanz zwischen dem lokalen Staat und der Lebenswelt der lokalen Bevölkerung.
Physical fitness is an important marker of health that enables people to carry out activities of daily living with vigour and alertness but without undue fatigue and with sufficient reserve to enjoy active leisure pursuits and to meet unforeseen emergencies. Especially, due to scientific findings that the onset of civilization diseases (e.g., obesity, cardiovascular disease) begins in childhood and that physical fitness tracks (at least) into young adulthood, the regular monitoring and promotion of physical fitness in children is risen up to a public health issue. In relation to the evaluation of a child’s physical fitness over time (i.e., development) the use of longitudinally-based percentile values is of particular interest due to their underlined dedication of true physical fitness development within subjects (i.e., individual changes in timing and tempo of growth and maturation). Besides its genetic determination (e.g., sex, body height), physical fitness is influenced by factors that refer to children’s environment and behaviour. For instance, disparities in physical fitness according to children’s living area are frequently reported concerning the fact that living in rural areas as compared to urban areas seems to be more favourable for children’s physical fitness. In addition, cross-sectional studies found higher fitness values in children participating in sports clubs as compared to non-participants. However, up to date, the observed associations between both (i.e., living area and sports club participating) and children’s physical fitness are unresolved concerning a long-term effect. In addition, social inequality as determined by the socioeconomic status (SES) extends through many areas of children’s life. While evidence indicates that the SES is inversely related to various indices of child’s daily life and behaviour like educational success, nutritional habits, and sedentary- and physical activity behaviour, a potential relationship between child’s physical fitness and the SES is hardly investigated and indicated inconsistent results.
The present thesis addressed three objectives: (1) to generate physical fitness percentiles for 9- to 12- year-old boys and girls using a longitudinal approach and to analyse the age- and sex-specific development of physical fitness, (2) to investigate the long-term effect of living area and sports club participation on physical fitness in third- to sixth-grade primary school students, and (3) to examine associations between the SES and physical fitness in a large and representative (i.e., for a German federal state) sample of third grade primary school students.
Methods
(i/ii) Healthy third graders were followed over four consecutive years (up to grade 6), including annually assessment of physical fitness and parental questionnaire (i.e., status of sports club participation and living area). Six tests were conducted to estimate various components of physical fitness: speed (50-m sprint test), upper body muscular power (1-kg ball push test), lower body muscular power (triple hop test), flexibility (stand-and-reach test), agility (star agility run test), and cardiorespiratory fitness (CRF) (9-min run test). (iii) Within a cross-sectional study (i.e., third objective), physical fitness of third graders was assessed by six physical fitness tests including: speed (20-m sprint test), upper body muscular power (1-kg ball push test), lower body muscular power (standing long jump [SLJ] test), flexibility (stand-and-reach test), agility (star agility run test), and CRF (6-min run test). By means of questionnaire, students reported their status of organized sports participation (OSP).
Results
(i) With respect to percentiles of physical fitness development, test performances increased in boys and girls from age 9 to 12, except for males’ flexibility (i.e., stable performance over time). Girls revealed significantly better performance in flexibility, whereas boys scored significantly higher in the remaining physical fitness tests. In girls as compared to boys, physical fitness development was slightly faster for upper body muscular power but substantially faster for flexibility. Generated physical fitness percentile curves indicated a timed and capacity-specific physical fitness development (curvilinear) for upper body muscular power, agility, and CRF. (ii) Concerning the effect of living area and sports club participation on physical fitness development, children living in urban areas showed a significantly faster performance development in physical fitness components of upper and lower body muscular power as compared to peers from rural areas. The same direction was noted as a trend in CRF. Additionally, children that regularly participated in a sports club, when compared to those that not continuously participated in a sports club demonstrated a significantly faster performance development in lower body muscular power. A trend of faster performance development in sports club participants occurred in CRF too. (iii) Regarding the association of SES with physical fitness, the percentage of third graders that achieved a high physical fitness level in lower body muscular power and CRF was significantly higher in students attending schools in communities with high SES as compared to middle and low SES, irrespective of sex. Similar, students from the high SES-group performed significantly better in lower body muscular power and CRF than students from the middle and/or the low SES-group.
Conclusion
(i) The generated percentile values provide an objective tool to estimate childrenʼs physical fitness within the frame of physical education (e.g., age- and sex-specific grading of motor performance) and further to detect children with specific fitness characteristics (low fit or high fit) that may be indicative for the necessity of preventive health promotion or long term athlete development. (ii) It is essential to consider variables of different domains (e.g., environment and behavior) in order to improve knowledge of potential factors which influence physical fitness during childhood. In this regard, the present thesis provide a first input to clarify the causality of living area and sports club participation on physical fitness development in school-aged children. Living in urban areas as well as a regular participation in sports clubs positively affected children´s physical fitness development (i.e., muscular power and CRF). Herein, sports club participation seems to be a key factor within the relationship between living area and physical fitness. (iii) The findings of the present thesis imply that attending schools in communities with high SES refers to better performance in specific physical fitness test items (i.e., muscular power, CRF) in third graders. Extra-curricular physical education classes may represent an important equalizing factor for physical activity opportunities in children of different SES backgrounds. In regard to strong evidence of a positive relationship between physical fitness - in particular muscular fitness/ CRF - and health, more emphasis should be laid on establishing sports clubs and extra-curricular physical education classes as an easy and attractive means to promote fitness-, and hence health- enhancing daily physical activity for all children (i.e. public health approach).
Analysis and modeling of transient earthquake patterns and their dependence on local stress regimes
(2015)
Investigations in the field of earthquake triggering and associated interactions, which includes aftershock triggering as well as induced seismicity, is important for seismic hazard assessment due to earthquakes destructive power. One of the approaches to study earthquake triggering and their interactions is the use of statistical earthquake models, which are based on knowledge of the basic seismicity properties, in particular, the magnitude distribution and spatiotemporal properties of the triggered events.
In my PhD thesis I focus on some specific aspects of aftershock properties, namely, the relative seismic moment release of the aftershocks with respect to the mainshocks; the spatial correlation between aftershock occurrence and fault deformation; and on the influence of aseismic transients on the aftershock parameter estimation. For the analysis of aftershock sequences I choose a statistical approach, in particular, the well known Epidemic Type Aftershock Sequence (ETAS) model, which accounts for the input of background and triggered seismicity. For my specific purposes, I develop two ETAS model modifications in collaboration with Sebastian Hainzl. By means of this approach, I estimate the statistical aftershock parameters and performed simulations of aftershock sequences as well.
In the case of seismic moment release of aftershocks, I focus on the ratio of cumulative seismic moment release with respect to the mainshocks. Specifically, I investigate the ratio with respect to the focal mechanism of the mainshock and estimate an effective magnitude, which represents the cumulative aftershock energy (similar to Bath's law, which defines the average difference between mainshock and the largest aftershock magnitudes). Furthermore, I compare the observed seismic moment ratios with the results of the ETAS simulations. In particular, I test a restricted ETAS (RETAS) model which is based on results of a clock advanced model and static stress triggering.
To analyze spatial variations of triggering parameters I focus in my second approach on the aftershock occurrence triggered by large mainshocks and the study of the aftershock parameter distribution and their spatial correlation with the coseismic/postseismic slip and interseismic locking. To invert the aftershock parameters I improve the modified ETAS (m-ETAS) model, which is able to take the extension of the mainshock rupture into account. I compare the results obtained by the classical approach with the output of the m-ETAS model.
My third approach is concerned with the temporal clustering of seismicity, which might not only be related to earthquake-earthquake interactions, but also to a time-dependent background rate, potentially biasing the parameter estimations. Thus, my coauthors and I also applied a modification of the ETAS model, which is able to take into account time-dependent background activity. It can be applicable for two different cases: when an aftershock catalog has a temporal incompleteness or when the background seismicity rate changes with time, due to presence of aseismic forces.
An essential part of any research is the testing of the developed models using observational data sets, which are appropriate for the particular study case. Therefore, in the case of seismic moment release I use the global seismicity catalog. For the spatial distribution of triggering parameters I exploit two aftershock sequences of the Mw8.8 2010 Maule (Chile) and Mw 9.0 2011 Tohoku (Japan) mainshocks. In addition, I use published geodetic slip models of different authors. To test our ability to detect aseismic transients my coauthors and I use the data sets from Western Bohemia (Central Europe) and California.
Our results indicate that:
(1) the seismic moment of aftershocks with respect to mainshocks depends on the static stress changes and is maximal for the normal, intermediate for thrust and minimal for strike-slip stress regimes, where the RETAS model shows a good correspondence with the results;
(2) The spatial distribution of aftershock parameters, obtained by the m-ETAS model, shows anomalous values in areas of reactivated crustal fault systems. In addition, the aftershock density is found to be correlated with coseismic slip gradient, afterslip, interseismic coupling and b-values. Aftershock seismic moment is positively correlated with the areas of maximum coseismic slip and interseismically locked areas. These correlations might be related to the stress level or to material properties variations in space;
(3) Ignoring aseismic transient forcing or temporal catalog incompleteness can lead to the significant under- or overestimation of the underlying trigger parameters. In the case when a catalog is complete, this method helps to identify aseismic sources.
In many procedures of seismic risk mitigation, ground motion simulations are needed to test systems or improve their effectiveness. For example they may be used to estimate the level of ground shaking caused by future earthquakes. Good physical models for ground motion simulation are also thought to be important for hazard assessment, as they could close gaps in the existing datasets. Since the observed ground motion in nature shows a certain variability, part of which cannot be explained by macroscopic parameters such as magnitude or position of an earthquake, it would be desirable that a good physical model is not only able to produce one single seismogram, but also to reveal this natural variability.
In this thesis, I develop a method to model realistic ground motions in a way that is computationally simple to handle, permitting multiple scenario simulations. I focus on two aspects of ground motion modelling. First, I use deterministic wave propagation for the whole frequency range – from static deformation to approximately 10 Hz – but account for source variability by implementing self-similar slip distributions and rough fault interfaces. Second, I scale the source spectrum so that the modelled waveforms represent the correct radiated seismic energy. With this scaling I verify whether the energy magnitude is suitable as an explanatory variable, which characterises the amount of energy radiated at high frequencies – the advantage of the energy magnitude being that it can be deduced from observations, even in real-time.
Applications of the developed method for the 2008 Wenchuan (China) earthquake, the 2003 Tokachi-Oki (Japan) earthquake and the 1994 Northridge (California, USA) earthquake show that the fine source discretisations combined with the small scale source variability ensure that high frequencies are satisfactorily introduced, justifying the deterministic wave propagation approach even at high frequencies. I demonstrate that the energy magnitude can be used to calibrate the high-frequency content in ground motion simulations.
Because deterministic wave propagation is applied to the whole frequency range, the simulation method permits the quantification of the variability in ground motion due to parametric uncertainties in the source description. A large number of scenario simulations for an M=6 earthquake show that the roughness of the source as well as the distribution of fault dislocations have a minor effect on the simulated variability by diminishing directivity effects, while hypocenter location and rupture velocity more strongly influence the variability. The uncertainty in energy magnitude, however, leads to the largest differences of ground motion amplitude between different events, resulting in a variability which is larger than the one observed.
For the presented approach, this dissertation shows (i) the verification of the computational correctness of the code, (ii) the ability to reproduce observed ground motions and (iii) the validation of the simulated ground motion variability. Those three steps are essential to evaluate the suitability of the method for means of seismic risk mitigation.
Methicillin resistant Staphylococcus aureus (MRSA) is one of the most important antibiotic-resistant pathogens in hospitals and the community. Recently, a new generation of MRSA, the so called livestock associated (LA) MRSA, has emerged occupying food producing animals as a new niche. LA-MRSA can be regularly isolated from economically important live-stock species including corresponding meats. The present thesis takes a methodological approach to confirm the hypothesis that LA-MRSA are transmitted along the pork, poultry and beef production chain from animals at farm to meat on consumers` table. Therefore two new concepts were developed, adapted to differing data sets.
A mathematical model of the pig slaughter process was developed which simulates the change in MRSA carcass prevalence during slaughter with special emphasis on identifying critical process steps for MRSA transmission. Based on prevalences as sole input variables the model framework is able to estimate the average value range of both the MRSA elimination and contamination rate of each of the slaughter steps. These rates are then used to set up a Monte Carlo simulation of the slaughter process chain. The model concludes that regardless of the initial extent of MRSA contamination low outcome prevalences ranging between 0.15 and 1.15 % can be achieved among carcasses at the end of slaughter. Thus, the model demonstrates that the standard procedure of pig slaughtering in principle includes process steps with the capacity to limit MRSA cross contamination. Scalding and singeing were identified as critical process steps for a significant reduction of superficial MRSA contamination.
In the course of the German national monitoring program for zoonotic agents MRSA prevalence and typing data are regularly collected covering the key steps of different food production chains. A new statistical approach has been proposed for analyzing this cross sectional set of MRSA data with regard to show potential farm to fork transmission. For this purpose, chi squared statistics was combined with the calculation of the Czekanowski similarity index to compare the distributions of strain specific characteristics between the samples from farm, carcasses after slaughter and meat at retail. The method was implemented on the turkey and veal production chains and the consistently high degrees of similarity which have been revealed between all sample pairs indicate MRSA transmission along the chain.
As the proposed methods are not specific to process chains or pathogens they offer a broad field of application and extend the spectrum of methods for bacterial transmission assessment.
In den letzten Jahrzehnten ist der Trend der Verselbstständigung in vielen Kommunen zu beobachten. Ein Großteil der öffentlichen Leistungserbringer wird mittlerweile als privatrechtliche Gesellschaften in einem wettbewerbsorientierten Umfeld geführt. Während viele Forscher Ausgliederungen in Form von nachgeordneten Behörden auf Bundesebene untersuchen und diese Reformwelle als einen faktischen Autonomisierungsprozess beschreiben, gibt es nur einige wenige Studien, die sich explizit mit den Autonomisierungstendenzen auf Kommunalebene auseinandersetzen. Daher fehlt es an empirischen Erkenntnissen zur Steuerung der kommunalen Beteiligungen.
In dieser Arbeit werden die Steuerungsarrangements deutscher Großstädte erstmals aus Sicht der Gesteuerten beleuchtet. Das Untersuchungsziel der vorliegenden Forschungsarbeit besteht darin, Flexibilisierungstendenzen in mehrheitlich kommunalen Unternehmen zu identifizieren und hierfür Erklärungsfaktoren zu identifizieren. Die Forschungsfrage lautet: Welche instrumentellen und relationalen Faktoren beeinflussen die Managementautonomie in kommunalen Mehrheitsbeteiligungen?
Dabei interessiert insbesondere die Einflussnahme der Kommunen auf verschiedene Tätigkeitsbereiche ihrer Ausgliederungen. Über diese unternehmensspezifischen Sachverhalte ist in Deutschland fast nichts und international nur sehr wenig Empirisches bekannt. Zur Beantwortung der Forschungsfrage hat der Autor auf Basis der Transaktionskosten- und der Social-Exchange-Theorie einen Analyserahmen erstellt. Die aufgestellten Hypothesen wurden mit einer großflächigen Umfrage bei 243 Unternehmen in den 39 größten deutschen Städten empirisch getestet.
Im Ergebnis zeigen sich mehrere empirische Erkenntnisse: Erstens konnten mittels Faktorenanalyse vier unabhängige Faktoren von Managementautonomie in kommunalen Unternehmen identifiziert werden: Personalautonomie, Generelles Management, Preisautonomie und Strategische Fragen. Während die Kommunen ihren Beteiligungen einen hohen Grad an Personalautonomie zugestehen, unterliegen vor allem strategische Investitionsentscheidungen wie die finanzielle Beteiligung an Tochterfirmen, große Projektvorhaben, Diversifikationsentscheidungen oder Kreditautfnahmen einem starken politischen Einfluss.
Zweitens führt eine Rechtsformänderung und die Platzierung in einem Wettbewerbsumfeld (auch bekannt als Corporatisation) vor allem zu einer größeren Flexibilisierung der Personal- und Preispolitik, wirkt sich allerdings wenig auf die weiteren Faktoren der Managementautonomie, Generelles Management und Strategische Entscheidungen, aus. Somit behalten die Kommunen ihre Möglichkeit, auf wichtige Unternehmensfragen der Beteiligung Einfluss zu nehmen, auch im Fall einer Formalprivatisierung bei.
Letztlich können zur Erklärung der Autonomiefaktoren transaktionskostenbasierte und relationale Faktoren ergänzend herangezogen werden. In den Transaktionsspezifika wirken vor allem der wahrgenommene Wettbewerb in der Branche, die Messbarkeit der Leistung, Branchenvariablen, die Anzahl der Politiker im Aufsichtsrat und die eingesetzten Steuerungsmechanismen. In den relationalen Faktoren setzen sich die Variablen gegenseitiges Vertrauen, Effektivität der Aufsichtsräte, Informationsaustausch, Rollenkonflikte, Rollenambivalenzen und Geschäftsführererfahrung im Sektor durch.
The rise of evolutionary novelties is one of the major drivers of evolutionary diversification. African weakly-electric fishes (Teleostei, Mormyridae) have undergone an outstanding adaptive radiation, putatively owing to their ability to communicate through species-specific Electric Organ Discharges (EODs) produced by a novel, muscle-derived electric organ. Indeed, such EODs might have acted as effective pre-zygotic isolation mechanisms, hence favoring ecological speciation in this group of fishes. Despite the evolutionary importance of this organ, genetic investigations regarding its origin and function have remained limited.
The ultimate aim of this study is to better understand the genetic basis of EOD production by exploring the transcriptomic profiles of the electric organ and of its ancestral counterpart, the skeletal muscle, in the genus Campylomormyrus. After having established a set of reference transcriptomes using “Next-Generation Sequencing” (NGS) technologies, I performed in silico analyses of differential expression, in order to identify sets of genes that might be responsible for the functional differences observed between these two kinds of tissues. The results of such analyses indicate that: i) the loss of contractile activity and the decoupling of the excitation-contraction processes are reflected by the down-regulation of the corresponding genes in the electric organ; ii) the metabolic activity of the electric organ might be specialized towards the production and turnover of membrane structures; iii) several ion channels are highly expressed in the electric organ in order to increase excitability, and iv) several myogenic factors might be down-regulated by transcription repressors in the EO.
A secondary task of this study is to improve the genus level phylogeny of Campylomormyrus by applying new methods of inference based on the multispecies coalescent model, in order to reduce the conflict among gene trees and to reconstruct a phylogenetic tree as closest as possible to the actual species-tree. By using 1 mitochondrial and 4 nuclear markers, I was able to resolve the phylogenetic relationships among most of the currently described Campylomormyrus species. Additionally, I applied several coalescent-based species delimitation methods, in order to test the hypothesis that putatively cryptic species, which are distinguishable only from their EOD, belong to independently evolving lineages. The results of this analysis were additionally validated by investigating patterns of diversification at 16 microsatellite loci. The results suggest the presence of a new, yet undescribed species of Campylomormyrus.
This thesis investigates temporal and aspectual reference in the typologically unrelated African languages Hausa (Chadic, Afro–Asiatic) and Medumba (Grassfields Bantu).
It argues that Hausa is a genuinely tenseless language and compares the interpretation of temporally unmarked sentences in Hausa to that of morphologically tenseless sentences in Medumba, where tense marking is optional and graded.
The empirical behavior of the optional temporal morphemes in Medumba motivates an analysis as existential quantifiers over times and thus provides new evidence suggesting that languages vary in whether their (past) tense is pronominal or quantificational (see also Sharvit 2014).
The thesis proposes for both Hausa and Medumba that the alleged future tense marker is a modal element that obligatorily combines with a prospective future shifter (which is covert in Medumba). Cross-linguistic variation in whether or not a future marker is compatible with non-future interpretation is proposed to be predictable from the aspectual architecture of the given language.
Reconstructing climate from the Dead Sea sediment record using high-resolution micro-facies analyses
(2015)
The sedimentary record of the Dead Sea is a key archive for reconstructing climate in the eastern Mediterranean region, as it stores the environmental and tectonic history of the Levant for the entire Quaternary. Moreover, the lake is located at the boundary between Mediterranean sub-humid to semi-arid and Saharo-Arabian hyper-arid climates, so that even small shifts in atmospheric circulation are sensitively recorded in the sediments. This DFG-funded doctoral project was carried out within the ICDP Dead Sea Deep Drilling Project (DSDDP) that intended to gain the first long, continuous and high-resolution sediment core from the deep Dead Sea basin. The drilling campaign was performed in winter 2010-11 and more than 700 m of sediments were recovered. The main aim of this thesis was (1) to establish the lithostratigraphic framework for the ~455 m long sediment core from the deep Dead Sea basin and (2) to apply high-resolution micro-facies analyses for reconstructing and better understanding climate variability from the Dead Sea sediments.
Addressing the first aim, the sedimentary facies of the ~455 m long deep-basin core 5017-1 were described in great detail and characterised through continuous overview-XRF element scanning and magnetic susceptibility measurements. Three facies groups were classified: (1) the marl facies group, (2) the halite facies group and (3) a group involving different expressions of massive, graded and slumped deposits including coarse clastic detritus. Core 5017-1 encompasses a succession of four main lithological units. Based on first radiocarbon and U-Th ages and correlation of these units to on-shore stratigraphic sections, the record comprises the last ca 220 ka, i.e. the upper part of the Amora Formation (parts of or entire penultimate interglacial and glacial), the last interglacial Samra Fm. (~135-75 ka), the last glacial Lisan Fm. (~75-14 ka) and the Holocene Ze’elim Formation. A major advancement of this record is that, for the first time, also transitional intervals were recovered that are missing in the exposed formations and that can now be studied in great detail.
Micro-facies analyses involve a combination of high-resolution microscopic thin section analysis and µXRF element scanning supported by magnetic susceptibility measurements. This approach allows identifying and characterising micro-facies types, detecting event layers and reconstructing past climate variability with up to seasonal resolution, given that the analysed sediments are annually laminated. Within this thesis, micro-facies analyses, supported by further sedimentological and geochemical analyses (grain size, X-ray diffraction, total organic carbon and calcium carbonate contents) and palynology, were applied for two time intervals:
(1) The early last glacial period ~117-75 ka was investigated focusing on millennial-scale hydroclimatic variations and lake level changes recorded in the sediments. Thereby, distinguishing six different micro-facies types with distinct geochemical and sedimentological characteristics allowed estimating relative lake level and water balance changes of the lake. Comparison of the results to other records in the Mediterranean region suggests a close link of the hydroclimate in the Levant to North Atlantic and Mediterranean climates during the time of the build-up of Northern hemisphere ice sheets during the early last glacial period.
(2) A mostly annually laminated late Holocene section (~3700-1700 cal yr BP) was analysed in unprecedented detail through a multi-proxy, inter-site correlation approach of a shallow-water core (DSEn) and its deep-basin counterpart (5017-1). Within this study, a ca 1500 years comprising time series of erosion and dust deposition events was established and anchored to the absolute time-scale through 14C dating and age modelling. A particular focus of this study was the characterisation of two dry periods, from ~3500 to 3300 and from ~3000 to 2400 cal yr BP, respectively. Thereby, a major outcome was the coincidence of the latter dry period with a period of moist and cold climate in Europe related to a Grand Solar Minimum around 2800 cal yr BP and an increase in flood events despite overall dry conditions in the Dead Sea region during that time. These contrasting climate signatures in Europe and at the Dead Sea were likely linked through complex teleconnections of atmospheric circulation, causing a change in synoptic weather patterns in the eastern Mediterranean.
In summary, within this doctorate the lithostratigraphic framework of a unique long sediment core from the deep Dead Sea basin is established, which serves as a base for any further high-resolution investigations on this core. It is demonstrated in two case studies that micro-facies analyses are an invaluable tool to understand the depositional processes in the Dead Sea and to decipher past climate variability in the Levant on millennial to seasonal time-scales. Hence, this work adds important knowledge helping to establish the deep Dead Sea record as a key climate archive of supra-regional significance.
Hintergrund: In Deutschland stellt der akute Myokardinfarkt (MI) eine der häufigsten Todesursachen dar. Als Ursache für regionale Unterschiede bei den Mortalitätsraten werden divergente Versorgungsstrukturen vermutet. Ziel der Untersuchung war, diese Fragestellung anhand anonymisierter krankenkassenbasierter Abrechnungsdaten zu evaluieren.
Methodik: Standardisierte Hospitalisierungs- sowie Krankenhaus- und Ein-Jahres-Mortalitätsraten nach MI wurden anhand anonymisierter Versichertendaten einer gesetzlichen Krankenkasse für das Jahr 2012 und die Bundesländer Berlin, Brandenburg und Mecklenburg-Vorpommern ermittelt (n=1.387.084, 46.3% male, 60.9 ± 18,2 years). Weiterhin wurden prädiktive Einflussfaktoren auf die Ein-Jahres-Mortalität, auf die Durchführung invasiver Prozeduren und auf eine leitliniengerechte pharmakotherapeutische Sekundärprävention analysiert.
Ergebnisse: 6.733 Patienten (73,7 ±13,0 Jahre, 56,7% männlich) wurden identifiziert. Obwohl für das Bundesland Berlin eine höhere Hospitalisierungsrate als in Mecklenburg-Vorpommern ermittelt werden konnte, ließen sich bei der Krankenhaus- und 1-Jahres-Mortalität keine signifikant abweichenden Raten zwischen den Bundesländern beobachten. Die Durchführung einer Koronarangiographie (OR: 0,42 [0,35-0,51]) und eine leitliniengerechte Pharmakotherapie (OR: 0,14 [0,12-0,17] waren mit einer geringeren 1-Jahres-Mortalität assoziiert. Die Durchführung einer Koronarangiographie und eine leitliniengerechte Pharmakotherapie von Patienten nach Myokardinfarkt wurde hingegen primär durch Alter und Geschlecht, nicht aber durch das Bundesland determiniert.
Folgerung: Eine regional divergierende stationäre und postinfarzielle Versorgung auf Bundesland-Ebene kann anhand der vorliegenden Daten nicht nachgewiesen werden.
Die vorliegende Dissertation analysiert die Mentalitäten der bürgerlichen Bewohner Bogotas 19. Jahrhundert. Einblicke in die Fremd- und Eigenperspektiven der Bogotaner werden durch die Analyse von Reiseliteratur gewonnen. Methodologisch stützt sich die Arbeit auf den Vergleich zwischen europäischen Reiseberichten aus dem 19. Jh., Chroniken aus dem 16. Jh. sowie zwei kolumbianische Romane aus dem frühen 20.. Jh. Die Texte werden historiographisch behandelt; obwohl sie unterschiedlichen literarischen Genres angehören, weisen sie einen gemeinsamen autobiographischen Charakter auf. Aus den Erfahrungen und Gedanken der Reisenden werden v.a. die Auswirkungen der geographischen und sozialen Isolation thematisiert, sowie die Einflüsse von politischen und religiösen Diskursen auf die Bildung von bürgerlichem Gedankengut.
The non-linear behaviour of the atmospheric dynamics is not well understood and makes the evaluation and usage of regional climate models (RCMs) difficult. Due to these non-linearities, chaos and internal variability (IV) within the RCMs are induced, leading to a sensitivity of RCMs to their initial conditions (IC). The IV is the ability of RCMs to realise different solutions of simulations that differ in their IC, but have the same lower and lateral boundary conditions (LBC), hence can be defined as the across-member spread between the ensemble members.
For the investigation of the IV and the dynamical and diabatic contributions generating the IV four ensembles of RCM simulations are performed with the atmospheric regional model HIRHAM5. The integration area is the Arctic and each ensemble consists of 20 members. The ensembles cover the time period from July to September for the years 2006, 2007, 2009 and 2012. The ensemble members have the same LBC and differ in their IC only. The different IC are arranged by an initialisation time that shifts successively by six hours. Within each ensemble the first simulation starts on 1st July at 00 UTC and the last simulation starts on 5th July at 18 UTC and each simulation runs until 30th September. The analysed time period ranges from 6th July to 30th September, the time period that is covered by all ensemble members. The model runs without any nudging to allow a free development of each simulation to get the full internal variability within the HIRHAM5.
As a measure of the model generated IV, the across-member standard deviation and the across-member variance is used and the dynamical and diabatic processes influencing the IV are estimated by applying a diagnostic budget study for the IV tendency of the potential temperature developed by Nikiema and Laprise [2010] and Nikiema and Laprise [2011]. The diagnostic budget study is based on the first law of thermodynamics for potential temperature and the mass-continuity equation. The resulting budget equation reveals seven contributions to the potential temperature IV tendency.
As a first study, this work analyses the IV within the HIRHAM5. Therefore, atmospheric circulation parameters and the potential temperature for all four ensemble years are investigated. Similar to previous studies, the IV fluctuates strongly in time. Further, due to the fact that all ensemble members are forced with the same LBC, the IV depends on the vertical level within the troposphere, with high values in the lower troposphere and at 500 hPa and low values in the upper troposphere and at the surface. By the same reason, the spatial distribution shows low values of IV at the boundaries of the model domain.
The diagnostic budget study for the IV tendency of potential temperature reveals that the seven contributions fluctuate in time like the IV. However, the individual terms reach different absolute magnitudes. The budget study identifies the horizontal and vertical ‘baroclinic’ terms as the main contributors to the IV tendency, with the horizontal ‘baroclinic’ term producing and the vertical ‘baroclinic’ term reducing the IV. The other terms fluctuate around zero, because they are small in general or are balanced due to the domain average.
The comparison of the results obtained for the four different ensembles (summers 2006, 2007, 2009 and 2012) reveals that on average the findings for each ensemble are quite similar concerning the magnitude and the general pattern of IV and its contributions. However, near the surface a weaker IV is produced with decreasing sea ice extent. This is caused by a smaller impact of the horizontal 'baroclinic' term over some regions and by the changing diabatic processes, particularly a more intense reducing tendency of the IV due to condensative heating. However, it has to be emphasised that the behaviour of the IV and its dynamical and diabatic contributions are influenced mainly by complex atmospheric feedbacks and large-scale processes and not by the sea ice distribution.
Additionally, a comparison with a second RCM covering the Arctic and using the same LBCs and IC is performed. For both models very similar results concerning the IV and its dynamical and diabatic contributions are found. Hence, this investigation leads to the conclusion that the IV is a natural phenomenon and is independent from the applied RCM.
The main goal of this cumulative thesis is the derivation of surface emissivity data in the infrared from radiance measurements of Venus. Since these data are diagnostic of the chemical composition and grain size of the surface material, they can help to improve knowledge of the planet’s geology. Spectrally resolved images of nightside emissions in the range 1.0-5.1 μm were recently acquired by the InfraRed Mapping channel of the Visible and InfraRed Thermal Imaging Spectrometer (VIRTIS-M-IR) aboard ESA’s Venus EXpress (VEX). Surface and deep atmospheric thermal emissions in this spectral range are strongly obscured by the extremely opaque atmosphere, but three narrow spectral windows at 1.02, 1.10, and 1.18 μm allow the sounding of the surface. Additional windows between 1.3 and 2.6 μm provide information on atmospheric parameters that is required to interpret the surface signals. Quantitative data on surface and atmosphere can be retrieved from the measured spectra by comparing them to simulated spectra. A numerical radiative transfer model is used in this work to simulate the observable radiation as a function of atmospheric, surface, and instrumental parameters. It is a line-by-line model taking into account thermal emissions by surface and atmosphere as well as absorption and multiple scattering by gases and clouds. The VIRTIS-M-IR measurements are first preprocessed to obtain an optimal data basis for the subsequent steps. In this process, a detailed detector responsivity analysis enables the optimization of the data consistency. The measurement data have a relatively low spectral information content, and different parameter vectors can describe the same measured spectrum equally well. A usual method to regularize the retrieval of the wanted parameters from a measured spectrum is to take into account a priori mean values and standard deviations of the parameters to be retrieved. This decreases the probability to obtain unreasonable parameter values. The multi-spectrum retrieval algorithm MSR is developed to additionally consider physically realistic spatial and temporal a priori correlations between retrieval parameters describing different measurements. Neglecting geologic activity, MSR also allows the retrieval of an emissivity map as a parameter vector that is common to several spectrally resolved images that cover the same surface target. Even applying MSR, it is difficult to obtain reliable emissivity maps in absolute values. A detailed retrieval error analysis based on synthetic spectra reveals that this is mainly due to interferences from parameters that cannot be derived from the spectra themselves, but that have to be set to assumed values to enable the radiative transfer simulations. The MSR retrieval of emissivity maps relative to a fixed emissivity is shown to effectively avoid most emissivity retrieval errors. Relative emissivity maps at 1.02, 1.10, and 1.18 μm are finally derived from many VIRTIS-M-IR measurements that cover a surface target at Themis Regio. They are interpreted as spatial variations relative to an assumed emissivity mean of the target. It is verified that the maps are largely independent of the choice of many interfering parameters as well as the utilized measurement data set. These are the first Venus IR emissivity data maps based on a consistent application of a full radiative transfer simulation and a retrieval algorithm that respects a priori information. The maps are sufficiently reliable for future geologic interpretations.
The Brazilian Cerrado is recognised as one of the most threatened biomes in the world, as the region has experienced a striking change from natural vegetation to intense cash crop production. The impacts of rapid agricultural expansion on soil and water resources are still poorly understood in the region. Therefore, the overall aim of the thesis is to improve our understanding of the ecohydrological processes causing water and soil degradation in the Brazilian Cerrado.
I first present a metaanalysis to provide quantitative evidence and identifying the main impacts of soil and water alterations resulting from land use change. Second, field studies were conducted to (i) examine the effects of land use change on soils of natural cerrado transformed to common croplands and pasture and (ii) indicate how agricultural production affects water quality across a meso-scale catchment. Third, the ecohydrological process-based model SWAT was tested with simple scenario analyses to gain insight into the impacts of land use and climate change on the water cycling in the upper São Lourenço catchment which experienced decreasing discharges in the last 40 years.
Soil and water quality parameters from different land uses were extracted from 89 soil and 18 water studies in different regions across the Cerrado. Significant effects on pH, bulk density and available P and K for croplands and less-pronounced effects on pastures were evident. Soil total N did not differ between land uses because most of the cropland sites were N-fixing soybean cultivations, which are not artificially fertilized with N. By contrast, water quality studies showed N enrichment in agricultural catchments, indicating fertilizer impacts and potential susceptibility to eutrophication. Regardless of the land use, P is widely absent because of the high-fixing capacities of deeply weathered soils and the filtering capacity of riparian vegetation. Pesticides, however, were consistently detected throughout the entire aquatic system. In several case studies, extremely high-peak concentrations exceeded Brazilian and EU water quality limits, which pose serious health risks.
My field study revealed that land conversion caused a significant reduction in infiltration rates near the soil surface of pasture (–96 %) and croplands (–90 % to –93 %). Soil aggregate stability was significantly reduced in croplands than in cerrado and pasture. Soybean crops had extremely high extractable P (80 mg kg–1), whereas pasture N levels declined. A snapshot water sampling showed strong seasonality in water quality parameters. Higher temperature, oxi-reduction potential (ORP), NO2–, and very low oxygen concentrations (<5 mg•l–1) and saturation (<60 %) were recorded during the rainy season. By contrast, remarkably high PO43– concentrations (up to 0.8 mg•l–1) were measured during the dry season. Water quality parameters were affected by agricultural activities at all sampled sub-catchments across the catchment, regardless of stream characteristic. Direct NO3– leaching appeared to play a minor role; however, water quality is affected by topsoil fertiliser inputs with impact on small low order streams and larger rivers. Land conversion leaving cropland soils more susceptible to surface erosion by increased overland flow events.
In a third study, the field data were used to parameterise SWAT. The model was tested with different input data and calibrated in SWAT-CUP using the SUFI-2 algorithm. The model was judged reliable to simulate the water balance in the Cerrado. A complete cerrado, pasture and cropland cover was used to analyse the impact of land use on water cycling as well as climate change projections (2039–2058) according to the projections of the RCP 8.5 scenario. The actual evapotranspiration (ET) for the cropland scenario was higher compared to the cerrado cover (+100 mm a–1). Land use change scenarios confirmed that deforestation caused higher annual ET rates explaining partly the trend of decreased streamflow. Taking all climate change scenarios into account, the most likely effect is a prolongation of the dry season (by about one month), with higher peak flows in the rainy season. Consequently, potential threats for crop production with lower soil moisture and increased erosion and sediment transport during the rainy season are likely and should be considered in adaption plans.
From the three studies of the thesis I conclude that land use intensification is likely to seriously limit the Cerrado’s future regarding both agricultural productivity and ecosystem stability. Because only limited data are available for the vast biome, we recommend further field studies to understand the interaction between terrestrial and aquatic systems. This thesis may serve as a valuable database for integrated modelling to investigate the impact of land use and climate change on soil and water resources and to test and develop mitigation measures for the Cerrado in the future.
Nowadays, business processes are increasingly supported by IT services that produce massive amounts of event data during process execution. Aiming at a better process understanding and improvement, this event data can be used to analyze processes using process mining techniques. Process models can be automatically discovered and the execution can be checked for conformance to specified behavior. Moreover, existing process models can be enhanced and annotated with valuable information, for example for performance analysis. While the maturity of process mining algorithms is increasing and more tools are entering the market, process mining projects still face the problem of different levels of abstraction when comparing events with modeled business activities. Mapping the recorded events to activities of a given process model is essential for conformance checking, annotation and understanding of process discovery results. Current approaches try to abstract from events in an automated way that does not capture the required domain knowledge to fit business activities. Such techniques can be a good way to quickly reduce complexity in process discovery. Yet, they fail to enable techniques like conformance checking or model annotation, and potentially create misleading process discovery results by not using the known business terminology.
In this thesis, we develop approaches that abstract an event log to the same level that is needed by the business. Typically, this abstraction level is defined by a given process model. Thus, the goal of this thesis is to match events from an event log to activities in a given process model. To accomplish this goal, behavioral and linguistic aspects of process models and event logs as well as domain knowledge captured in existing process documentation are taken into account to build semiautomatic matching approaches. The approaches establish a pre--processing for every available process mining technique that produces or annotates a process model, thereby reducing the manual effort for process analysts. While each of the presented approaches can be used in isolation, we also introduce a general framework for the integration of different matching approaches.
The approaches have been evaluated in case studies with industry and using a large industry process model collection and simulated event logs. The evaluation demonstrates the effectiveness and efficiency of the approaches and their robustness towards nonconforming execution logs.
Business process management (BPM) is a systematic and structured approach to model, analyze, control, and execute business operations also referred to as business processes that get carried out to achieve business goals. Central to BPM are conceptual models. Most prominently, process models describe which tasks are to be executed by whom utilizing which information to reach a business goal. Process models generally cover the perspectives of control flow, resource, data flow, and information systems.
Execution of business processes leads to the work actually being carried out. Automating them increases the efficiency and is usually supported by process engines. This, though, requires the coverage of control flow, resource assignments, and process data. While the first two perspectives are well supported in current process engines, data handling needs to be implemented and maintained manually. However, model-driven data handling promises to ease implementation, reduces the error-proneness through graphical visualization, and reduces development efforts through code generation.
This thesis addresses the modeling, analysis, and execution of data in business processes and presents a novel approach to execute data-annotated process models entirely model-driven. As a first step and formal grounding for the process execution, a conceptual framework for the integration of processes and data is introduced. This framework is complemented by operational semantics through a Petri net mapping extended with data considerations. Model-driven data execution comprises the handling of complex data dependencies, process data, and data exchange in case of communication between multiple process participants. This thesis introduces concepts from the database domain into BPM to enable the distinction of data operations, to specify relations between data objects of the same as well as of different types, to correlate modeled data nodes as well as received messages to the correct run-time process instances, and to generate messages for inter-process communication. The underlying approach, which is not limited to a particular process description language, has been implemented as proof-of-concept.
Automation of data handling in business processes requires data-annotated and correct process models. Targeting the former, algorithms are introduced to extract information about data nodes, their states, and data dependencies from control information and to annotate the process model accordingly. Usually, not all required information can be extracted from control flow information, since some data manipulations are not specified. This requires further refinement of the process model. Given a set of object life cycles specifying allowed data manipulations, automated refinement of the process model towards containment of all data manipulations is enabled. Process models are an abstraction focusing on specific aspects in detail, e.g., the control flow and the data flow views are often represented through activity-centric and object-centric process models. This thesis introduces algorithms for roundtrip transformations enabling the stakeholder to add information to the process model in the view being most appropriate.
Targeting process model correctness, this thesis introduces the notion of weak conformance that checks for consistency between given object life cycles and the process model such that the process model may only utilize data manipulations specified directly or indirectly in an object life cycle. The notion is computed via soundness checking of a hybrid representation integrating control flow and data flow correctness checking. Making a process model executable, identified violations must be corrected. Therefore, an approach is proposed that identifies for each violation multiple, alternative changes to the process model or the object life cycles.
Utilizing the results of this thesis, business processes can be executed entirely model-driven from the data perspective in addition to the control flow and resource perspectives already supported before. Thereby, the model creation is supported by algorithms partly automating the creation process while model consistency is ensured by data correctness checks.
Continental rifts are excellent regions where the interplay between extension, the build-up of topography, erosion and sedimentation can be evaluated in the context of landscape evolution. Rift basins also constitute important archives that potentially record the evolution and migration of species and the change of sedimentary conditions as a result of climatic change. Finally, rifts have increasingly become targets of resource exploration, such as hydrocarbons or geothermal systems. The study of extensional processes and the factors that further modify the mainly climate-driven surface process regime helps to identify changes in past and present tectonic and geomorphic processes that are ultimately recorded in rift landscapes.
The Cenozoic East African Rift System (EARS) is an exemplary continental rift system and ideal natural laboratory to observe such interactions. The eastern and western branches of the EARS constitute first-order tectonic and topographic features in East Africa, which exert a profound influence on the evolution of topography, the distribution and amount of rainfall, and thus the efficiency of surface processes. The Kenya Rift is an integral part of the eastern branch of the EARS and is characterized by high-relief rift escarpments bounded by normal faults, gently tilted rift shoulders, and volcanic centers along the rift axis.
Considering the Cenozoic tectonic processes in the Kenya Rift, the tectonically controlled cooling history of rift shoulders, the subsidence history of rift basins, and the sedimentation along and across the rift, may help to elucidate the morphotectonic evolution of this extensional province. While tectonic forcing of surface processes may play a minor role in the low-strain rift on centennial to millennial timescales, it may be hypothesized that erosion and sedimentation processes impacted by climate shifts associated with pronounced changes in the availability in moisture may have left important imprints in the landscape.
In this thesis I combined thermochronological, geomorphic field observations, and morphometry of digital elevation models to reconstruct exhumation processes and erosion rates, as well as the effects of climate on the erosion processes in different sectors of the rift. I present three sets of results: (1) new thermochronological data from the northern and central parts of the rift to quantitatively constrain the Tertiary exhumation and thermal evolution of the Kenya Rift. (2) 10Be-derived catchment-wide mean denudation rates from the northern, central and southern rift that characterize erosional processes on millennial to present-day timescales; and (3) paleo-denudation rates in the northern rift to constrain climatically controlled shifts in paleoenvironmental conditions during the early Holocene (African Humid Period).
Taken together, my studies show that time-temperature histories derived from apatite fission track (AFT) analysis, zircon (U-Th)/He dating, and thermal modeling bracket the onset of rifting in the Kenya Rift between 65-50 Ma and about 15 Ma to the present. These two episodes are marked by rapid exhumation and, uplift of the rift shoulders. Between 45 and 15 Ma the margins of the rift experienced very slow erosion/exhumation, with the accommodation of sediments in the rift basin.
In addition, I determined that present-day denudation rates in sparsely vegetated parts of the Kenya Rift amount to 0.13 mm/yr, whereas denudation rates in humid and more densely vegetated sectors of the rift flanks reach a maximum of 0.08 mm/yr, despite steeper hillslopes. I inferred that hillslope gradient and vegetation cover control most of the variation in denudation rates across the Kenya Rift today. Importantly, my results support the notion that vegetation cover plays a fundamental role in determining the voracity of erosion of hillslopes through its stabilizing effects on the land surface.
Finally, in a pilot study I highlighted how paleo-denudation rates in climatic threshold areas changed significantly during times of transient hydrologic conditions and involved a sixfold increase in erosion rates during increased humidity. This assessment is based on cosmogenic nuclide (10Be) dating of quartzitic deltaic sands that were deposited in the northern Kenya Rift during a highstand of Lake Suguta, which was associated with the Holocene African Humid Period. Taken together, my new results document the role of climate variability in erosion processes that impact climatic threshold environments, which may provide a template for potential future impacts of climate-driven changes in surface processes in the course of Global Change.
Spots on stellar surfaces are thought to be stellar analogues of sunspots. Thus, starspots are direct manifestations of strong magnetic fields. Their decay rate is directly related to the magnetic diffusivity, which itself is a key quantity for the deduction of an activity cycle length. So far, no single starspot decay has been observed, and thus no stellar activity cycle was inferred from its corresponding turbulent diffusivity.
We investigate the evolution of starspots on the rapidly-rotating K0 giant XX Triangulum. Continuous high-resolution and phase-resolved spectroscopy was obtained with the robotic 1.2-m STELLA telescope on Tenerife over a timespan of six years. With our line-profile inversion code iMap we reconstruct a total of 36 consecutive Doppler maps. To quantify starspot area decay and growth, we match the observed images with simplified spot models based on a Monte-Carlo approach.
It is shown that the surface of XX Tri is covered with large high-latitude and even polar spots and with occasional small equatorial spots. Just over the course of six years, we see a systematically changing spot distribution with various time scales and morphology such as spot fragmentation and spot merging as well as spot decay and formation.
For the first time, a starspot decay rate on another star than the Sun is determined. From our spot-decay analysis we determine an average linear decay rate of D = -0.067±0.006 Gm^2/day. From this decay rate, we infer a turbulent diffusivity of η_τ = (6.3±0.5) x 10^14 cm^2/s and consequently predict an activity cycle of 26±6 years. The obtained cycle length matches very well with photometric observations.
Our time-series of Doppler maps further enables to investigate the differential rotation of XX Tri. We therefore applied a cross-correlation analysis. We detect a weak solar-like differential rotation with a surface shear of α = 0.016±0.003. This value agrees with similar studies of other RS CVn stars.
Furthermore, we found evidence for active longitudes and flip-flops. Whereas the more active longitude is located in phase towards the (unseen) companion star, the weaker active longitude is located at the opposite stellar hemisphere. From their periodic appearance, we infer a flip-flop cycle of ~2 years. Both activity phenomena are common on late-type binary stars.
Last but not least we redetermine several astrophysical properties of XX Tri and its binary system, as large datasets of photometric and spectroscopic observations are available since its last determination in 1999. Additionally, we compare the rotational spot-modulation from photometric and spectroscopic studies.
The Barberton Greenstone Belt (BGB) in the northwestern part of South Africa belongs to the few well-preserved remnants of Archean crust. Over the last centuries, the BGB has been intensively studied at surface with detailed mapping of its surfacial geological units and tectonic features. Nevertheless, the deeper structure of the BGB remains poorly understood. Various tectonic evolution models have been developed based on geo-chronological and structural data. These theories are highly controversial and centre on the question whether plate tectonics - as geoscientists understand them today - was already evolving on the Early Earth or whether vertical mass movements driven by the higher temperature of the Earth in Archean times governed continent development.
To get a step closer to answering the questions regarding the internal structure and formation of the BGB, magnetotelluric (MT) field experiments were conducted as part of the German-South African research initiative Inkaba yeAfrica. Five-component MT data (three magnetic and two electric channels) were collected at ~200 sites aligned along six profiles crossing the southern part of the BGB. Tectonic features like (fossil) faults and shear zones are often mineralized and therefore can have high electrical conductivities. Hence, by obtaining an image of the conductivity distribution of the subsurface from MT measurements can provide useful information on tectonic processes.
Unfortunately, the BGB MT data set is heavily affected by man-made electromagnetic noise caused, e.g. by powerlines and electric fences. Aperiodic spikes in the magnetic and corresponding offsets in the electric field components impair the data quality particularly at periods >1 s which are required to image deep electrical structures. Application of common methods for noise reduction like delay filtering and remote reference processing, only worked well for periods <1 s. Within the framework of this thesis two new filtering approaches were developed to handle the severe noise in long period data and obtain reliable processing results. The first algorithm is based on the Wiener filter in combination with a spike detection algorithm. Comparison of data variances of a local site with those of a reference site allows the identification of disturbed time series windows for each recorded channel at the local site. Using the data of the reference site, a Wiener filter algorithm is applied to predict physically meaningful data to replace the disturbed windows. While spikes in the magnetic channels are easily recognized and replaced, steps in the electric channels are more difficult to detect depending on their offset. Therefore, I have implemented a novel approach based on time series differentiation, noise removal and subsequent integration to overcome this obstacle. A second filtering approach where spikes and steps in the time series are identified using a comparison of the short and long time average of the data was also implemented as part of my thesis. For this filtering approach the noise in the form of spikes and offsets in the data is treated by an interpolation of the affected data samples. The new developments resulted in a substantial data improvement and allowed to gain one to two decades of data (up to 10 or 100 s).
The re-processed MT data were used to image the electrical conductivity distribution of the BGB by 2D and 3D inversion. Inversion models are in good agreement with the surface geology delineating the highly resistive rocks of the BGB from surrounding more conductive geological units. Fault zones appear as conductive structures and can be traced to depths of 5 to 10 km. 2D models suggest a continuation of the faults further south across the boundary of the BGB. Based on the shallow tectonic structures (fault system) within the BGB compared to deeply rooted resistive batholiths in the area, tectonic models including both vertical mass transport and in parts present-day style plate tectonics seem to be most likely for the evolution of the BGB.
Spectral fingerprinting
(2015)
Current research on runoff and erosion processes, as well as an increasing demand for sustainable watershed management emphasize the need for an improved understanding of sediment dynamics. This involves the accurate assessment of erosion rates and sediment transfer, yield and origin. A variety of methods exist to capture these processes at the catchment scale. Among these, sediment fingerprinting, a technique to trace back the origin of sediment, has attracted increasing attention by the scientific community in recent years. It is a two-step procedure, based on the fundamental assumptions that potential sources of sediment can be reliably discriminated based on a set of characteristic ‘fingerprint’ properties, and that a comparison of source and sediment fingerprints allows to quantify the relative contribution of each source.
This thesis aims at further assessing the potential of spectroscopy to assist and improve the sediment fingerprinting technique. Specifically, this work focuses on (1) whether potential sediment sources can be reliably identified based on spectral features (‘fingerprints’), whether (2) these spectral fingerprints permit the quantification of relative source contribution, and whether (3) in situ derived source information is sufficient for this purpose. Furthermore, sediment fingerprinting using spectral information is applied in a study catchment to (4) identify major sources and observe how relative source contributions change between and within individual flood events. And finally, (5) spectral fingerprinting results are compared and combined with simultaneous sediment flux measurements to study sediment origin, transport and storage behaviour.
For the sediment fingerprinting approach, soil samples were collected from potential sediment sources within the Isábena catchment, a meso-scale basin in the central Spanish Pyrenees. Undisturbed samples of the upper soil layer were measured in situ using an ASD spectroradiometer and subsequently sampled for measurements in the laboratory. Suspended sediment was sampled automatically by means of ISCO samplers at the catchment as well as at the five major subcatchment outlets during flood events, and stored fine sediment from the channel bed was collected from 14 cross-sections along the main river. Artificial mixtures of known contributions were produced from source soil samples. Then, all source, sediment and mixture samples were dried and spectrally measured in the laboratory. Subsequently, colour coefficients and physically based features with relation to organic carbon, iron oxide, clay content and carbonate, were calculated from all in situ and laboratory spectra. Spectral parameters passing a number of prerequisite tests were submitted to principal component analyses to study natural clustering of samples, discriminant function analyses to observe source differentiation accuracy, and a mixing model for source contribution assessment. In addition, annual as well as flood event based suspended sediment fluxes from the catchment and its subcatchments were calculated from rainfall, water discharge and suspended sediment concentration measurements using rating curves and Quantile Regression Forests. Results of sediment flux monitoring were interpreted individually with respect to storage behaviour, compared to fingerprinting source ascriptions and combined with fingerprinting to assess their joint explanatory potential.
In response to the key questions of this work, (1) three source types (land use) and five spatial sources (subcatchments) could be reliably discriminated based on spectral fingerprints. The artificial mixture experiment revealed that while (2) laboratory parameters permitted source contribution assessment, (3) the use of in situ derived information was insufficient. Apparently, high discrimination accuracy does not necessarily imply good quantification results. When applied to suspended sediment samples of the catchment outlet, the spectral fingerprinting approach was able to (4) quantify the major sediment sources: badlands and the Villacarli subcatchment, respectively, were identified as main contributors, which is consistent with field observations and previous studies. Thereby, source contribution was found to vary both, within and between individual flood events. Also sediment flux was found to vary considerably, annually as well as seasonally and on flood event base. Storage was confirmed to play an important role in the sediment dynamics of the studied catchment, whereas floods with lower total sediment yield tend to deposit and floods with higher yield rather remove material from the channel bed. Finally, a comparison of flux measurements with fingerprinting results highlighted the fact that (5) immediate transport from sources to the catchment outlet cannot be assumed. A combination of the two methods revealed different aspects of sediment dynamics that none of the techniques could have uncovered individually.
In summary, spectral properties provide a fast, non-destructive, and cost-efficient means to discriminate and quantify sediment sources, whereas, unfortunately, straight-forward in situ collected source information is insufficient for the approach. Mixture modelling using artificial mixtures permits valuable insights into the capabilities and limitations of the method and similar experiments are strongly recommended to be performed in the future. Furthermore, a combination of techniques such as e.g. (spectral) sediment fingerprinting and sediment flux monitoring can provide comprehensive understanding of sediment dynamics.
Microsaccades
(2015)
The first thing we do upon waking is open our eyes. Rotating them in our eye sockets, we scan our surroundings and collect the information into a picture in our head. Eye movements can be split into saccades and fixational eye movements, which occur when we attempt to fixate our gaze. The latter consists of microsaccades, drift and tremor. Before we even lift our eye lids, eye movements – such as saccades and microsaccades that let the eyes jump from one to another position – have partially been prepared in the brain stem. Saccades and microsaccades are often assumed to be generated by the same mechanisms. But how saccades and microsaccades can be classified according to shape has not yet been reported in a statistical manner. Research has put more effort into the investigations of microsaccades’ properties and generation only since the last decade. Consequently, we are only beginning to understand the dynamic processes governing microsaccadic eye movements. Within this thesis, the dynamics governing the generation of microsaccades is assessed and the development of a model for the underlying processes. Eye movement trajectories from different experiments are used, recorded with a video-based eye tracking technique, and a novel method is proposed for the scale-invariant detection of saccades (events of large amplitude) and microsaccades (events of small amplitude). Using a time-frequency approach, the method is examined with different experiments and validated against simulated data. A shape model is suggested that allows for a simple estimation of saccade- and microsaccade related properties. For sequences of microsaccades, in this thesis a time-dynamic Markov model is proposed, with a memory horizon that changes over time and which can best describe sequences of microsaccades.
Intuitively, it is clear that neural processes and eye movements in reading are closely connected, but only few studies have investigated both signals simultaneously. Instead, the usual approach is to record them in separate experiments and to subsequently consolidate the results. However, studies using this approach have shown that it is feasible to coregister eye movements and EEG in natural reading and contributed greatly to the understanding of oculomotor processes in reading. The present thesis builds upon that work, assessing to what extent coregistration can be helpful for sentence processing research.
In the first study, we explore how well coregistration is suited to study subtle effects common to psycholinguistic experiments by investigating the effect of distance on dependency resolution. The results demonstrate that researchers must improve the signal-to-noise ratio to uncover more subdued effects in coregistration. In the second study, we compare oscillatory responses in different presentation modes. Using robust effects from world knowledge violations, we show that the generation and retrieval of memory traces may differ between natural reading and word-by-word presentation. In the third study, we bridge the gap between our knowledge of behavioral and neural responses to integration difficulties in reading by analyzing the EEG in the context of regressive saccades. We find the P600, a neural indicator of recovery processes, when readers make a regressive saccade in response to integration difficulties.
The results in the present thesis demonstrate that coregistration can be a useful tool for the study of sentence processing. However, they also show that it may not be suitable for some questions, especially if they involve subtle effects.
Im Rahmen der EU-weiten REACH-Verordnung haben Alternativmethoden zum Tierversuch in der Toxikologie an Bedeutung gewonnen. Die Alternativmethoden gliedern sich auf in In-vitro- und In-silico-Methoden. In dieser Dissertation wurden verschiedene Konzepte der In-silico-Toxikologie behandelt.
Die bearbeiteten Themen reichen von quantitativen Strukturaktivitätsbeziehungen (QSAR) über eine neue Herangehensweise an das gängige Konzept zur Festlegung von Grenzwerten bis hin zu computerbasierten Modellierungen zum Alkohol- und Bisphenol-A-Stoffwechsel.
Das Kapitel über QSAR befasst sich im Wesentlichen mit der Erstellung und Analyse einer Datenbank mit 878 Substanzen, die sich aus Tierversuchsstudien aus dem Archiv des Bundesinstituts für Risikobewertung zusammensetzt. Das Design wurde dabei an eine bereits bestehende Datenbank angepasst, um so einen möglichst großen Datenpool zu generieren. In der Analyse konnte u.a. gezeigt werden, dass Stoffe mit niedrigerem Molekulargewicht ein erhöhtes Potential für toxikologische Schäden aufwiesen als größere Moleküle.
Mit Hilfe des sogenannten TTC-Konzepts können Grenzwerte für Stoffe geringer Exposition festgelegt werden, zu denen keine toxikologischen Daten zur Verfügung stehen. In dieser Arbeit wurden für die Stoffe dreier Datenbanken entsprechende Grenzwerte festgelegt. Es erfolgte zunächst eine gängige strukturbasierte Aufteilung der Substanzen in die Kategorien "nicht toxisch", "möglicherweise toxisch" und "eindeutig toxisch". Substanzen, die aufgrund ihrer Struktur in eine der drei Klassen eingeordnet werden, erhalten den entsprechenden Grenzwert. Da in die dritte Klasse auch Stoffe eingeordnet werden, deren Toxizität nicht bestimmbar ist, ist sie sehr groß. Daher wurden in dieser Arbeit die ersten beiden Klassen zusammengelgt, um einen größeren Datenpool zu ermöglichen. Eine weitere Neuerung umfasst die Erstellung eines internen Grenzwerts. Diese Vorgehensweise hat den Vorteil, dass der Expositionsweg herausgerechnet wird und somit beispielsweise Studien mit oraler Verabreichung mit Studien dermaler Verabreichung verglichen werden können.
Mittels physiologisch basiertem kinetischem Modelling ist es möglich, Vorgänge im menschlichen Körper mit Hilfe spezieller Software nachzuvollziehen. Durch diese Vorgehensweise können Expositionen von Chemikalien simuliert werden. In einem Teil der Arbeit wurden Alkoholexpositionen von gestillten Neugeborenen simuliert, deren Mütter unmittelbar zuvor alkoholische Getränke konsumiert hatten. Mit dem Modell konnte gezeigt werden, dass die Expositionen des Kindes durchweg gering waren. Nach einem Glas Wein wurden Spitzenkonzentrationen im Blut von Neugeborenen von 0,0034 Promille ermittelt. Zum Vergleich wurde die Exposition durch ein für Säuglinge zugelassenes alkoholhaltiges pflanzliches Arzneimittel simuliert. Hier wurden Spitzenkonzentrationen von 0,0141 Promille erreicht. Daher scheinen Empfehlungen wie gelegentlicher Konsum ohne schädigende Wirkung auf das Kind wissenschaftlich fundiert zu sein.
Ein weiteres Kinetik-Modell befasste sich mit dem Stoffwechsel von Bisphenol A. Teils widersprüchliche Daten zur Belastung mit BPA in der wissenschaftlichen Literatur führen wiederholt zu Anregungen, den Grenzwert der Chemikalie anzupassen. Die Funktionalität der am Metabolismus beteiligten Enzyme kann je nach Individuum unterschiedlich ausgeprägt sein. Mittels Modellings konnte hier gezeigt werden, dass dies maßgeblich dazu führt, dass sich berechnete Plasmaspiegel von Individuen bis zu 4,7-fach unterscheiden.
Die Arbeit konnte somit einen Beitrag zur Nutzung und Weiterentwicklung von In-silico-Modellen für diverse toxikologische Fragestellungen leisten.
Stream water and groundwater are important fresh water resources but their water quality is deteriorated by harmful solutes introduced by human activities. The interface between stream water and the subsurface water is an important zone for retention, transformation and attenuation of these solutes. Streambed structures enhance these processes by increased water and solute exchange across this interface, denoted as hyporheic exchange.
This thesis investigates the influence of hydrological and morphological factors on hyporheic water and solute exchange as well as redox-reactions in fluvial streambed structures on the intermediate scale (10–30m). For this purpose, a three-dimensional numerical modeling approach for coupling stream water flow with porous media flow is used. Multiple steady state stream water flow scenarios over different generic pool-riffle morphologies and a natural in-stream gravel bar are simulated by a computational fluid dynamics code that provides the hydraulic head distribution at the streambed. These heads are subsequently used as the top boundary condition of a reactive transport groundwater model of the subsurface beneath the streambed. Ambient groundwater that naturally interacts with the stream water is considered in scenarios of different magnitudes of downwelling stream water (losing case) and upwelling groundwater (gaining case). Also, the neutral case, where stream stage and groundwater levels are balanced is considered. Transport of oxygen, nitrate and dissolved organic carbon and their reaction by aerobic respiration and denitrification are modeled.
The results show that stream stage and discharge primarily induce hyporheic exchange flux and solute transport with implications for specific residence times and reactions at both the fully and partially submerged structures. Gaining and losing conditions significantly diminish the extent of the hyporheic zone, the water exchange flux, and shorten residence times for both the fully and partially submerged structures. With increasing magnitude of gaining or losing conditions, these metrics exponentially decrease.
Stream water solutes are transported mainly advectively into the hyporheic zone and hence their influx corresponds directly to the infiltrating water flux. Aerobic respiration takes place in the shallow streambed sediments, coinciding to large parts with the extent of the hyporheic exchange flow. Denitrification occurs mainly as a “reactive fringe” surrounding the aerobic zone, where oxygen concentration is low and still a sufficient amount of stream water carbon source is available. The solute consumption rates and the efficiency of the aerobic and anaerobic reactions depend primarily on the available reactive areas and the residence times, which are both controlled by the interplay between hydraulic head distribution at the streambed and the gradients between stream stage and ambient groundwater. Highest solute consumption rates can be expected under neutral conditions, where highest solute flux, longest residence times and largest extent of the hyporheic exchange occur. The results of this thesis show that streambed structures on the intermediate scale have a significant potential to contribute to a net solute turnover that can support a healthy status of the aquatic ecosystem.
We study segregation of the subducted oceanic crust (OC) at the core mantle boundary and its ability to accumulate and form large thermochemical piles (such as the seismically observed Large Low Shear Velocity Provinces - LLSVPs). Our high-resolution numerical simulations suggest that the longevity of LLSVPs for up to three billion years, and possibly longer, can be ensured by a balance in the rate of segregation of high-density OC-material to the CMB, and the rate of its entrainment away from the CMB by mantle upwellings.
For a range of parameters tested in this study, a large-scale compositional anomaly forms at the CMB, similar in shape and size to the LLSVPs. Neutrally buoyant thermochemical piles formed by mechanical stirring - where thermally induced negative density anomaly is balanced by the presence of a fraction of dense anomalous material - best resemble the geometry of LLSVPs. Such neutrally buoyant piles tend to emerge and survive for at least 3Gyr in simulations with quite different parameters. We conclude that for a plausible range of values of density anomaly of OC material in the lower mantle - it is likely that it segregates to the CMB, gets mechanically mixed with the ambient material, and forms neutrally buoyant large scale compositional anomalies similar in shape to the LLSVPs.
We have developed an efficient FEM code with dynamically adaptive time and space resolution, and marker-in-cell methodology. This enabled us to model thermochemical mantle convection at realistically high convective vigor, strong thermally induced viscosity variations, and long term evolution of compositional fields.
In this thesis we study reciprocal classes of Markov chains. Given a continuous time Markov chain on a countable state space, acting as reference dynamics, the associated reciprocal class is the set of all probability measures on path space that can be written as a mixture of its bridges. These processes possess a conditional independence property that generalizes the Markov property, and evolved from an idea of Schrödinger, who wanted to obtain a probabilistic interpretation of quantum mechanics.
Associated to a reciprocal class is a set of reciprocal characteristics, which are space-time functions that determine the reciprocal class. We compute explicitly these characteristics, and divide them into two main families: arc characteristics and cycle characteristics. As a byproduct, we obtain an explicit criterion to check when two different Markov chains share their bridges.
Starting from the characteristics we offer two different descriptions of the reciprocal class, including its non-Markov probabilities.
The first one is based on a pathwise approach and the second one on short time asymptotic. With the first approach one produces a family of functional equations whose only solutions are precisely the elements of the reciprocal class. These equations are integration by parts on path space associated with derivative operators which perturb the paths by mean of the addition of random loops. Several geometrical tools are employed to construct such formulas. The problem of obtaining sharp characterizations is also considered, showing some interesting connections with discrete geometry. Examples of such formulas are given in the framework of counting processes and random walks on Abelian groups, where the set of loops has a group structure.
In addition to this global description, we propose a second approach by looking at the short time behavior of a reciprocal process. In the same way as the Markov property and short time expansions of transition probabilities characterize Markov chains, we show that a reciprocal class is characterized by imposing the reciprocal property and two families of short time expansions for the bridges. Such local approach is suitable to study reciprocal processes on general countable graphs. As application of our characterization, we considered several interesting graphs, such as lattices, planar
graphs, the complete graph, and the hypercube.
Finally, we obtain some first results about concentration of measure implied by lower bounds on the reciprocal characteristics.
The relationship between nutrition and the development of chronic diseases including metabolic syndrome, diabetes mellitus, cancer and cardiovascular disease has been well studied. On the other hand, changes in the GH-IGF-1 axis in association with nutrition-related diseases have been reported. The interplay between GH, total IGF-1 and different inhibitory and stimulatory kinds of IGF-1 binding proteins (IGFBPs) results in IGF-1 bioactivity, the ability of IGF-1 to induce phosphorylation of its receptor and consequently its signaling. Moreover, IGF-1 bioactivity is sufficient to reflect any change in the GH-IGF-1 system. Accumulating evidence suggests that both of high protein diet, characterized by increased glucagon secretion, and insulin-induced hypoglycemia increase mortality rate and the mechanisms are unclear. However both of glucagon and insulin-induced hypoglycemia are potent stimuli of GH secretion. The aim of the current study was to identify the impact of glucagon and insulin-induced hypoglycemia on IGF-1 bioactivity as possible mechanisms. In a double-blind placebo-controlled study, glucagon was intramuscularly administrated in 13 type 1 diabetic patients (6 males /7 females; [BMI]: 24.8 ± 0.95 kg/m2), 11 obese subjects (OP; 5/ 6; 34.4 ± 1.7 kg/m2), and 13 healthy lean participants (LP; 6/ 7; 21.7 ± 0.6 kg/m2), whereas 12 obese subjects (OP; 6/ 6; 34.4 ± 1.7 kg/m2), and 13 healthy lean participants (LP; 6/ 7; 21.7 ± 0.6 kg/m2) performed insulin tolerance test in another double-blind placebo-controlled study and changes in GH, total IGF-1, IGF binding proteins (IGFBPs) and IGF-1 bioactivity, measured by the cell-based KIRA method, were investigated. In addition, the interaction between the metabolic hormones (glucagon and insulin) and the GH-IGF-1 system on the transcriptional level was studied using mouse primary hepatocytes. In this thesis, glucagon decreased IGF-1 bioactivity in humans independently of endogenous insulin levels, most likely through modulation of IGFBP-1 and-2 levels. The glucagon-induced reduction in IGF-1 bioactivity may represent a novel mechanism underlying the impact of glucagon on GH secretion and may explain the negative effect of high protein diet related to increased cardiovascular risk and mortality rate. In addition, insulin-induced hypoglycemia was correlated with a decrease in IGF-1 bioactivity through up-regulation of IGFBP-2. These results may refer to a possible and poorly explored mechanism explaining the strong association between hypoglycemia and increased cardiovascular mortality among diabetic patients.
This dissertation investigates the working memory mechanism subserving human sentence processing and its relative contribution to processing difficulty as compared to syntactic prediction. Within the last decades, evidence for a content-addressable memory system underlying human cognition in general has accumulated (e.g., Anderson et al., 2004). In sentence processing research, it has been proposed that this general content-addressable architecture is also used for language processing (e.g., McElree, 2000).
Although there is a growing body of evidence from various kinds of linguistic dependencies that is consistent with a general content-addressable memory subserving sentence processing (e.g., McElree et al., 2003; VanDyke2006), the case of reflexive-antecedent dependencies has challenged this view. It has been proposed that in the processing of reflexive-antecedent dependencies, a syntactic-structure based memory access is used rather than cue-based retrieval within a content-addressable framework (e.g., Sturt, 2003).
Two eye-tracking experiments on Chinese reflexives were designed to tease apart accounts assuming a syntactic-structure based memory access mechanism from cue-based retrieval (implemented in ACT-R as proposed by Lewis and Vasishth (2005).
In both experiments, interference effects were observed from noun phrases which syntactically do not qualify as the reflexive's antecedent but match the animacy requirement the reflexive imposes on its antecedent. These results are interpreted as evidence against a purely syntactic-structure based memory access. However, the exact pattern of effects observed in the data is only partially compatible with the Lewis and Vasishth cue-based parsing model.
Therefore, an extension of the Lewis and Vasishth model is proposed. Two principles are added to the original model, namely 'cue confusion' and 'distractor prominence'.
Although interference effects are generally interpreted in favor of a content-addressable memory architecture, an alternative explanation for interference effects in reflexive processing has been proposed which, crucially, might reconcile interference effects with a structure-based account.
It has been argued that interference effects do not necessarily reflect cue-based retrieval interference in a content-addressable memory but might equally well be accounted for by interference effects which have already occurred at the moment of encoding the antecedent in memory (Dillon, 2011).
Three experiments (eye-tracking and self-paced reading) on German reflexives and Swedish possessives were designed to tease apart cue-based retrieval interference from encoding interference. The results of all three experiments suggest that there is no evidence that encoding interference affects the retrieval of a reflexive's antecedent.
Taken together, these findings suggest that the processing of reflexives can be explained with the same cue-based retrieval mechanism that has been invoked to explain syntactic dependency resolution in a range of other structures. This supports the view that the language processing system is located within a general cognitive architecture, with a general-purpose content-addressable working memory system operating on linguistic expressions.
Finally, two experiments (self-paced reading and eye-tracking) using Chinese relative clauses were conducted to determine the relative contribution to sentence processing difficulty of working-memory processes as compared to syntactic prediction during incremental parsing.
Chinese has the cross-linguistically rare property of being a language with subject-verb-object word order and pre-nominal relative clauses. This property leads to opposing predictions of expectation-based
accounts and memory-based accounts with respect to the relative processing difficulty of subject vs. object relatives.
Previous studies showed contradictory results, which has been attributed to different kinds local ambiguities confounding the materials (Lin and Bever, 2011). The two experiments presented are the first to compare Chinese relatives clauses in syntactically unambiguous contexts.
The results of both experiments were consistent with the predictions of the expectation-based account of sentence processing but not with the memory-based account. From these findings, I conclude that any theory of human sentence processing needs to take into account the power of predictive processes unfolding in the human mind.
Ziel dieser Arbeit war die Synthese und Charakterisierung von neuartigen fluoreszierenden Copolymeren zur Analytdetektion in wässrigen Systemen. Das Detektionssystem sollte ein einfaches Schalten der Fluoreszenz bei Analytbindung „Aus“ bzw. Verdrängung „An“ ermöglichen. Dafür wurde die Synthese eines funktionalisierten Monomers so geplant, dass sich Fluorophor und Analyt innerhalb derselben Monomereinheit in direkter Nachbarschaft zueinander befinden. So sollten bei Erkennung des Analyten durch eine mit einem Fluoreszenzlöscher funktionalisierte Erkennungsstruktur Fluorophor und Löscher in einen vorgegebenen Abstand zueinander gezwungen und die Fluoreszenz des Fluorophors effizient gelöscht werden. Bei anschließender Verdrängung der Erkennungseinheit durch einen stärker bindenden Analyten sollte die Fluoreszenz wieder „angeschaltet“ werden. Eine weitere Zielstellung für das Detektionssystem war eine hohe Löslichkeit und Fluoreszenzintensität in Wasser. Da die Anwendung solcher Sensoren besonders in der Medizin und Biologie, z.B. für Schnellerkennungstest von Pathogenen, von Interesse ist, ist die Kompatibilität mit wässrigen Medien essentiell. Die funktionalisierten Monomere wurden frei radikalisch mit N Vinyl-pyrrolidon bzw. N Vinyl¬caprolactam zu wasserlöslichen, fluoreszierenden Copolymeren umgesetzt. In den N-Vinyl¬pyrrolidon-Polymeren (PNVP) wurde RhodaminB, in den thermoresponsiven N Vinyl¬caprolactam-Polymeren (PNVCL) ein Naphthalsäureimid als Fluorophor verwendet. Während Rhodamine eine hohe Fluoreszenzintensität, gute Quantenausbeuten und hohen Extinktionskoeffizienten in Wasser zeigen, sind Naphthalsäure¬imide umgebungssensitive Chromophore, die bei Änderung ihrer Lösungsmittelumgebung, wie z.B. beim Kollaps eines thermoresponsiven Polymers in Wasser, ihre Fluoreszenzintensität und Quantenausbeute drastisch ändern können. Der Vorteil der hier verwendeten Strategie der Monomersynthese liegt darin, dass bei jeder spezifischen Analytdetektion durch eine Erkennungseinheit die Fluoreszenz effizient gelöscht bzw. bei Verdrängung durch einen stärker bindenden Analyten wieder „angeschaltet“ wird. Dieses Prinzip wird bereits vielfach in der Biologie in sogenannten „Molecular Beacons“ ausgenutzt, wobei ein Fluorophor und ein Löscher durch spezifische DNA Basenpaarung in einen vorgegebenen Abstand zueinander gezwungen werden und so ein „Schalten“ der Fluoreszenz ermöglichen. Aufgrund der vorgegebenen Struktur der DNA Basensequenzen ist es jedoch nicht direkt auf andere Erkennungsreaktionen übertragbar. Daher wurde ein Modellsystem entwickelt, welches die Möglichkeit bietet Analyt, Erkennungseinheit und Signalgeber variabel, je nach Anforderungen des Systems, auszutauschen. So soll es möglich sein, den Sensor a priori für jede Erkennungs¬reaktion zu verwenden. Als Modell Bindungs¬paare wurden ß Cyclodextrin/Adamantan und Con¬cana¬valinA/Mannose ausgewählt. Adamantan bzw. Mannose wurde als Analyt zusammen mit dem Fluorophor in das Polymer eingebunden. ß Cyclo¬dextrin (ß CD) bzw. ConcanavalinA (ConA) wurde als Erkennungsstruktur an einem Fluoreszenzlöscher immobilisiert. Polymer-basierte Fluoreszenzsensoren sind in der Fachliteratur gut dokumentiert. In der Regel sind Signalgeber und Analyt jedoch statistisch im Polymer verteilt, da sie sich entweder in unterschiedlichen Monomereinheiten befinden oder die Funktionalisierung durch eine polymeranaloge Umsetzung erfolgt. Der gewählte Ansatz Fluorophor und Analyt innerhalb derselben Monomereinheit einzubinden, soll bei jeder Erkennungsreaktion des Analyten zu einer Änderung der Signalintensität des Fluorophors führen. Eine hohe Signalintensität bei Analytdetektion ist wünschenswert, insbesondere für Erkennungsreaktionen, die mit möglichst geringem apparativem Aufwand, am besten mit dem bloßen Auge zu verfolgen sein sollen. Des Weiteren ist es möglich den Fluorophorgehalt im Polymer genau einzustellen und so Selbstlöschung zu vermeiden. Die synthetisierten Polymere haben einen Fluorophorgehalt von 0,01 mol% bis 0,5 mol%. Für die RhodaminB haltigen Polymere zeigte sich, dass ein Fluorophorgehalt unterhalb 0,1 mol% im Polymer die höchsten Ausbeuten, Molmassen und Quantenausbeuten liefert. Für die Naphthalsäureimid haltigen Polymere hingegen wurden auch für einen Fluorophorgehalt von bis zu 1 mol% hohe Ausbeuten und Molmassen erreicht. Die Naphthalsäureimid haltigen Polymere haben jedoch in wässriger Lösungsmittelumgebung nur geringe Quantenausbeuten. Als Fluoreszenzlöscher wurden Goldnanopartikel synthetisiert, die mit den entsprechenden Erkennungsstrukturen (ß-CD oder ConA) für den verwendeten Analyten funktionalisiert wurden. Goldnanopartikel als Löscher bieten den Vorteil, dass ihre Dispergierbarkeit in einem Lösemittel durch Funktionalisierung ihrer Hülle gezielt gesteuert werden kann. Durch die hohe Affinität von Goldnanopartikeln zu Thiolen und Aminen konnten sie mit Hilfe einfacher Syntheseschritte mit Thio ß CD Derivaten bzw. ConA funktionalisiert werden. In der hier vorgelegten Arbeit sollte ein Modellsystem für einen solches fluoreszenz-basiertes Detektionssystem in Wasser entwickelt werden. Nachfolgend werden die zu erfüllenden strukturellen Voraussetzungen für die Synthese eines solchen Sensors nochmals zusammengefasst:
1. Verwendung eines Fluorophors, der eine hohe Signalintensität zeigt.
2. Analyt bzw. Erkennungseinheit soll sich im Abstand von wenigen Nanometern zum Signalgeber befinden, um bei jeder Detektionsreaktion die Signalintensität des Signalgebers beeinflussen zu können.
3. Die Detektionseinheit benötigt eine funktionelle Gruppe zur Immobilisierung. Immobilisierung kann z.B. durch Einbindung in ein Polymer erfolgen.
4. Der Fluorophor soll bei Änderung seiner lokalen Umgebung, durch Binden eines Löschers oder Änderung seiner Lösemittelumgebung seine Fluoreszenzeigenschaften drastisch ändern.
5. Die Reaktion sollte schnell und mit möglichst geringem apparativem Aufwand, am besten mit bloßem Auge zu verfolgen sein.
Für das ß-CD/Adamantan Modellsystem wurde ein Fluoreszenz Aus/An Sensor entwickelt, der bei Binden ß CD funktionalisierter Goldnanopartikel an das polymergebundene Adamantan die Fluoreszenz des RhodaminB Fluorophors effizient löscht und bei Verdrängung der Goldnanopartikel wieder zurück gewinnt. Dies konnte auch mit bloßem Auge verfolgt werden.
Für die Naphthalsäureimid Monomere, die mit NVCL copolymerisiert wurden, wurde abhängig von der lokalen Umgebung des Fluorophors eine unterschiedliche Verstärkung der Fluoreszenzintensität bei Überschreiten des Trübungspunktes des Polymers gefunden. Dabei zeigte sich, dass die Einführung eines Abstandshalters zwischen Polymerrückgrat und Fluorophor zu einer großen Fluoreszenz¬verstärkung führt, während sich ohne Abstandshalter die Fluoreszenzintensität bei Über¬schreiten des Trübungspunktes kaum ändert.
Two of the most controversial issues concerning the late Cenozoic evolution of the Andean orogen are the timing of uplift of the intraorogenic Puna plateau and its eastern border, the Eastern Cordillera, and ensuing changes in climatic and surface-process conditions in the intermontane basins of the NW-Argentine Andes. The Eastern Cordillera separates the internally drained, arid Puna from semi-arid intermontane basins and the humid sectors of the Andean broken foreland and the Subandean fold-and-thrust belt to the east. With elevations between 4,000 and 6,000 m the eastern flanks of the Andes form an efficient orographic barrier with westward-increasing elevation and asymmetric rainfall distribution and amount with respect to easterly moisture-bearing winds. This is mirrored by pronounced gradients in the efficiency of surface processes that erode and re-distribute sediment from the uplifting ranges. Although the overall pattern of deformation and uplift in this sector of the southern central Andes shows an eastward migration of deformation, a well-developed deformation front does not exist and uplift and associated erosion and sedimentary processes are highly disparate in space and time. In addition, periodic deformation within intermontane basins, and continued diachronous foreland uplifts associated with the reactivation of inherited basement structures furthermore make a rigorous assessment of the spatiotemporal uplift patterns difficult.
This thesis focuses on the tectonic evolution of the Eastern Cordillera of NW Argentina, the depositional history of its intermontane sedimentary basins, and the regional topographic evolution of the eastern flank of the Puna Plateau. The intermontane basins of the Eastern Cordillera and the adjacent morphotectonic provinces of the Sierras Pampeanas and the Santa Bárbara System are akin to reverse fault bounded, filled, and partly coalesced sedimentary basins of the Puna Plateau. In contrast to the Puna basins, however, which still form intact morphologic entities, repeated deformation, erosion, and re-filling have impacted the basins in the Eastern Cordillera. This has resulted in a rich stratigraphy of repeated basin fills, but many of these basins have retained vestiges of their early depositional history that may reach back in time when these areas were still part of a contiguous and undeformed foreland basin. Fortunately, these strata also contain abundant volcanic ashes that are not only important horizons to decipher tectono-sedimentary events through U-Pb geochronology and geochemical correlation, but they also represent terrestrial recorders of the hydrogen-isotope composition of ancient meteoric waters that can be compared to the isotopic composition of modern meteoric water. The ash horizons are thus unique recorders of past environmental conditions and lend themselves to tracking the development of rainfall barriers and tectonically forced climate and environmental change through time.
U-Pb zircon geochronology and paleocurrent reconstructions of conglomerate sequences in the Humahuaca Basin of the Eastern Cordillera at 23.5° S suggest that the basin was an integral part of a largely unrestricted depositional system until 4.2 Ma, which subsequently became progressively decoupled from the foreland by range uplifts to the east that forced easterly moisture-bearing winds to precipitate in increasingly eastward locations. Multiple cycles of severed hydrological conditions and drainage re-capture are identified together with these processes that were associated with basin filling and sediment evacuation, respectively. Moreover, systematic relationships among faults, regional unconformities and deformed landforms reveal a general pattern of intra-basin deformation that appears to be linked with basin-internal deformation during or subsequent to episodes of large-scale sediment removal. Some of these observations are supported by variations in the hydrogen stable isotope composition of volcanic glass from the Neogene to Quaternary sedimentary record, which can be related to spatiotemporal changes in topography and associated orographic effects. δDg values in the basin strata reveal two main trends associated with surface uplift in the catchment area between 6.0 and 3.5 Ma and the onset of semiarid conditions in the basin following the attainment of threshold elevations for effective orographic barriers to the east after 3.5 Ma. The disruption of sediment supply from western sources after 4.2 Ma and subsequent hinterland aridification, moreover, emphasize the possibility that these processes were related to lateral orogenic growth of the adjacent Puna Plateau. As a result of the hinterland aridification the regions in the orogen interior have been characterized by an inefficient fluvial system, which in turn has helped maintaining internal drainage conditions, sediment storage, and relief reduction within high-elevation basins.
The diachronous nature of basin formation and impacts on the fluvial system in the adjacent broken foreland is underscored by the results of detailed sediment provenance and paleocurrent analyses, as well as U-Pb zircon geochronology in the Lerma and Metán basins at ca. 25° S. This is particularly demonstrated by the isolated uplift of the Metán range at ~10 Ma, which is more than 50 km away from the presently active orogenic front along the eastern Puna margin and the Eastern Cordillera to the west. At about 5 Ma, Puna-sourced sediments disappear from the foreland record, documenting further range uplifts in the Eastern Cordillera and hydrological isolation of the neighboring Angastaco Basin from the foreland. Finally, during the late Pliocene and Quaternary, deformation has been accommodated across the entire foreland and is still active. To elucidate the interactions between tectonically controlled changes in elevation and their impact on atmospheric circulation processes in this region, this thesis provides additional, temporally well-constrained hydrogen stable isotope results of volcanic glass samples from the broken foreland, including the Angastaco Basin, and other intermontane basins farther south. The results suggest similar elevations of intermontane basins and the foreland sectors prior to ca. 7 Ma. In case of the Angastaco Basin the region was affected by km-scale surface uplift of the basin. A comparison with coeval isotope data collected from sedimentary sequences in the Puna plateau explains rapid shifts in the intermontane δDg record and supports the notion of recurring phases of enhanced deep convection during the Pliocene, and thus climatic conditions during the middle to late Pliocene similar to the present day.
Combined, field-based and isotope geochemical methods used in this study of the NW-Argentine Andes have thus helped to gain insight into the systematics, rate changes, interactions, and temporal characteristics among tectonically controlled deformation patterns, the build-up of topography impacting atmospheric processes, the distribution of rainfall, and resulting surface processes in a tectonically active mountain belt. Ultimately, this information is essential for a better understanding of the style and the rates at which non-collisional mountain belts evolve, including the development orogenic plateaus and their bordering flanks. The results presented in this study emphasize the importance of stable isotope records for paleoaltimetric and paleoenvironmental studies in mountain belts and furnishes important data for a rigorous interpretation of such records.
Analphabetismus und Teilhabe
(2015)
Aus bildungstheoretisch-gesellschaftskritischer Perspektive stellt sich Lernen als soziales Handeln in gesellschaftlich-vermittelten Verhältnissen – Möglichkeiten wie auch Begrenzungen – dar. Funktionaler Analphabetismus ist mit einem bundesweiten Anteil von 14% der erwerbsfähigen Bevölkerung oder 7,5 Millionen Analphabeten in Deutschland nicht nur ein bildungspolitisches und -praktisches, sondern auch ein wissenschaftlich zu untersuchendes Phänomen. Es gibt zahlreiche Untersuchungen, die sich mit dieser Thematik auseinandersetzen und Anknüpfungspunkte für die vorliegende Studie bieten. Aus der Zielgruppenforschung beispielsweise ist bekannt, dass die Hauptadressaten der Männer, der Älteren und der Bildungsfernen nicht adäquat erreicht bzw. als Teilnehmende gewonnen werden. Aus der Teilnehmendenforschung sind Abbrüche und Drop-Outs bekannt.
Warum Analphabeten im Erwachsenenalter, also nach der Aneignung vielfältigster Bewältigungsstrategien, durch das sich das Phänomen einer direkten Sichtbarkeit entzieht, dennoch beginnen das Lesen und Schreiben (wieder) zu lernen, wird bislang weder bildungs- noch lerntheoretisch untersucht. Im Rahmen der vorliegenden Erwachsenenbildungsstudie werden genau diese Lernanlässe empirisch herausgearbeitet.
Als Heuristik wird auf eine subjekttheoretische Theoriefolie rekurriert, die sich in besonderer Weise eignet Lernbegründungen im Kontext gesellschaftlich verhafteter Biografien sichtbar zu machen. Lernforschung im Begründungsmodell muss dabei auf eine Methodik zurückgreifen, die die Perspektive des Subjekts, Bedeutungszusammenhänge und typische Sinnstrukturen hervorbringen kann. Daher wird ein auf Einzelfallstudien basierendes, qualitatives Forschungsdesign gewählt, das Daten aus der Erhebung mittels problemzentrierter Interviews bereitstellt, die eine Auswertung innerhalb der Forschungsstrategie der Grounded Theory erfahren und in einer empirisch begründeten Typenbildung münden. Dieses Design ermöglicht die Rekonstruktion typischer Lernanlässe und im Ergebnis die Entwicklung einer gegenstandsbezogene Theorie mittlerer Reichweite.
Aus der vorliegenden Bedeutungs-Begründungsanalyse konnten empirisch fünf Lernbegründungstypen ausdifferenziert werden, die sich im Spannungsverhältnis von Teilhabeausrichtung und Widersprüchlichkeit bewegen und in ihrer Komplexität mittels der drei Schlüsselkategorien Bedeutungsraum, Reflexion der sozialen Eingebundenheit und Kompetenzen sowie Lernen bzw. dem Erleben der Diskrepanzerfahrung zwischen Lesen-Wollen und Lesen-Können dargestellt werden. Das Spektrum der Lernbegründungstypen reicht von teilhabesicherndem resignierten Lernen, bei dem die Sicherung des bedrohten Status quo im Vordergrund steht und die Welt als nicht gestaltbar erlebt wird, bis hin zu vielschichtigem teilhabeerweiternden Lernen, das auf die Erweiterung der eigenen Handlungsmöglichkeiten zielt und die umfangreichste Reflexion der sozialen Eingebundenheit und Kompetenzen aufweist. Funktionale Analphabeten begründen ihr Lernen und Nicht-Lernen vor dem Hintergrund ihrer sozialen Situation, ihrer Begrenzungen und Möglichkeiten: Schriftsprachlernen erhält erst im Kontext gesellschaftlicher Teilhabe und dessen Reflexion eine Bedeutung.
Mit der Einordnung der Lernbegründungen funktionaler Analphabeten in: erstens, Diskurse der Bildungsbenachteiligung durch Exklusionsprozesse; zweitens, die lerntheoretische Bedeutung von Inklusionsprozessen und drittens, den internationalen Theorieansatz transformativen Lernens durch die Integration der Reflexionskategorie, erfolgt eine Erweiterung bildungs- und lerntheoretischer Ansätze. In dieser Arbeit werden Alphabetisierungs- und Erwachsenen-bildungsforschung verbunden und in den jeweiligen Diskurs integriert. Weitere Anschluss- und Verwertungsmöglichkeiten in der Bildungsforschung wären denkbar. Die Untersuchung von Lernbegründungen im Längsschnitt beispielsweise kann Transformationsprozesse rekonstruierbar machen und somit Erträge für eine Bildungsprozessforschung liefern. Bildungspraktisch können die Lernbegründungstypen einerseits der Teilnehmergewinnung dienen, andererseits Ausgangspunkt für reflexive Lernbegleitungskonzepte sein, die Lernbegründungen zur Sprache bringen und die soziale Eingebundenheit thematisieren und damit Lernprozesse unterstützen.
The main research question of this thesis concerns the relation between focus interpretation, focus realization, and association with focus in the West Chadic language Ngamo.
Concerning the relation between focus realization and interpretation, this thesis contributes to the question, cross-linguistically, what factors influence a marked realization of the focus/background distinction. There is background-marking rather than focus-marking in Ngamo, and the background marker is related to the definite determiner in the language. Using original fieldwork data as a basis, a formal semantic analysis of the background marker as a definite determiner of situations is proposed.
Concerning the relation between focus and association with focus, the thesis adds to the growing body of crosslinguistic evidence that not all so-called focus-sensitive operators always associate with focus. The thesis shows that while the exclusive particle yak('i) (= "only") in Ngamo conventionally associates
with focus, the particles har('i) (= "even, as far as, until, already"), and ke('e) (= "also, and") do not.
The thesis provides an analysis of these phenomena in a situation semantic framework.
In this thesis we utilize resolved stellar populations to improve our understanding of galaxy formation and evolution. In the first part we improve a method for metallicity determination of faint old stellar systems, in the second and third part we analyze the individual history of six nearby disk galaxies outside the Local Group.
A New Calibration of the Color Metallicity Relation of Red Giants for HST data:
It is well known, that the color distribution of stars on the the Red Giant Branch (RGB) can be used to determine metallicities of old stellar populations that have only shallow photometry. Based on the largest sample of globular clusters ever used for such studies, we quantify the relation between metallicity and color in the widely used HST ACS filters F606W and F814W.
We use a sample of globular clusters from the ACS Globular Cluster Survey and measure their RGB color at given absolute magnitudes to derive the color-metallicity relation. We find a clear relation between metallicity and RGB color; we investigate the scatter and the uncertainties in this relation and show its limitations. A comparison with isochrones shows reasonably good agreement with BaSTI models, a small offset to Dartmouth models, and a larger offset to Padua models.
Even for the best globular cluster data available, the metallicity of a simple stellar population can be determined from the RGB alone only with an accuracy of 0.3 dex for [M/H]<-1, and 0.15 dex for [M/H]>-1. For mixed populations, as they are observed in external galaxies, the uncertainties will be even larger due to uncertainties in extinction, age, etc. Therefore caution is necessary when interpreting photometric metallicities.
The Structural History of Nearby Low Mass Disk Galaxies:
We study the individual evolution histories of three nearby, low-mass, edge-on galaxies (IC5052, NGC4244, NGC5023).
Using the color magnitude diagrams of resolved stellar populations, we construct star count density maps for populations of different ages and analyze the change of structural parameters with stellar age within each galaxy.
The three galaxies show low vertical heating rates, which are much lower than the heating rate of the Milky Way. This indicates that heating agents, as giant molecular clouds and spiral structure are weak in low mass galaxies.
We do not detect a separate thick disk in any of the three galaxies, even though our observations cover a larger range in equivalent surface brightness than any integrated light study. While scaleheights increase with age, each population can be well described by a single disk. Only two of the galaxies contain a very weak additional component, which we identify as the faint halo. The mass of these faint halos is less than 1% of the mass of the disk.
All populations in the three galaxies exhibit no or only little flaring. While this finding is consistent with previous integrated light studies, it poses strong constraints on galaxy formation models, because most theoretical simulations often find strong flaring due to interactions or radial migration.
Furthermore, we find breaks in the radial profiles of all three galaxies. The radii of these breaks are independent of age, and the break strength is decreasing with age in two of the galaxies (NGC4244 and NGC5023). This is consistent with break formation models, that combine a star formation cutoff with radial migration. The differing behavior of IC5052 can be explained by a recent interaction or minor merger.
The Structural History of Massive Disk Galaxies:
We extend the structural analysis of stellar populations with distinct ages to three massive galaxies, NGC891, NGC4565 and NGC7814. While confusion effects due to the high stellar number densities in their central region, and the prominent dust lanes inhibit an detailed analysis of the radial profiles, we can study their vertical structure.
These massive galaxies also have a slower heating than the Milky Way, comparable to the low mass galaxies. This can be traced back to their already thick young populations and thick layers of their interstellar medium.
We do not find a clear separate thick disk in any of these three galaxies; all populations can be described by a single disk plus a S\'ersic bulge/halo component. In contrast to the low mass galaxies, we cannot rule out the presence of thick disks in the massive galaxies, because of the strong influence of the halo, that might hide the possible contribution of the thick disk to the vertical star count profiles. However, the faintness of the possible thick disks still points to problems in the earlier ubiquitous findings of thick disks in external galaxies.
Die Honigbiene Apis mellifera zeigt innerhalb einer Kolonie eine an das Alter gekoppelte Arbeitsteilung. Junge Honigbienen versorgen die Brut (Ammenbienen), während ältere Honigbienen (Sammlerinnen) außerhalb des Stocks Pollen und Nektar eintragen. Die biogenen Amine Octopamin und Tyramin sind an der Steuerung der Arbeitsteilung maßgeblich beteiligt. Sie interagieren mit Zielzellen über die Bindung an G Protein gekoppelte Rezeptoren. A. mellifera besitzt fünf charakterisierte Octopaminrezeptoren (AmOctαR1, AmOctβR1-4), einen charakterisierten Tyraminrezeptor (AmTyr1) sowie einen weiteren putativen Tyraminrezeptor.
In der vorliegenden Arbeit wurde dieser putative Aminrezeptor als zweiter Tyraminrezeptor (AmTyr2) identifiziert, lokalisiert und pharmakologisch charakterisiert.
Die von der cDNA abgeleitete Aminosäuresequenz weist strukturelle Eigenschaften und konservierte Motive von G Protein gekoppelten Rezeptoren auf. Phylogenetisch ordnet sich der AmTyr2 Rezeptor bei den Tyramin 2 Rezeptoren anderer Insekten ein. Die funktionelle und pharmakologische Charakterisierung des putativen Tyraminrezeptors erfolgte in modifizierten HEK293 Zellen, die mit der Rezeptor cDNA transfiziert wurden. Die Applikation von Tyramin aktiviert Adenylylcyclasen in diesen Zellen und resultiert in einem Anstieg des intrazellulären cAMP Gehalts. Der AmTyr2 Rezeptor kann durch Tyramin in nanomolaren Konzentrationen halbmaximal aktiviert werden. Während es sich bei Octopamin um einen wirkungsvollen Agonisten des Rezeptors handelt, sind Mianserin und Yohimbin effektive Antagonisten. Für die Lokalisierung des Rezeptorproteins wurde ein polyklonaler Antikörper generiert. Eine AmTyr2-ähnliche Immunreaktivität zeigt sich im Gehirn in den optischen Loben, den Antennalloben, dem Zentralkomplex und in den Kenyon Zellen der Pilzkörper.
Des Weiteren wurde die Rolle der Octopamin- und Tyraminrezeptoren bei der Steuerung der altersabhängigen Arbeitsteilung analysiert.
Die Genexpression des AmOctαR1 in verschiedenen Gehirnteilen korreliert unabhängig vom Alter mit der sozialen Rolle, während sich die Genexpression von AmOctβR3/4 und den Tyraminrezeptoren AmTyr1 und AmTyr2 maximal mit dem Alter aber nicht der sozialen Rolle ändert. Sammlerinnen weisen einen höheren Octopamingehalt im Gesamtgehirn auf als Ammenbienen; bei Tyramin zeigen sich keine Unterschiede. Während Tyramin offensichtlich keine direkte Rolle spielt, werden durch Octopamin gesteuerte Prozesse der altersabhängigen Arbeitsteilung bei der Honigbiene vermutlich über den AmOctαR1 vermittelt.
Die Ergebnisse der vorliegenden Arbeit zeigen die wichtige Rolle von biogenen Aminen, insbesondere Octopamin bei der sozialen Organisation von Insektenstaaten.
Die Dissertation mit dem Thema „Demografie und politischer Reduktionismus – eine Diskursanalyse der Demografiepolitik in Deutschland“ knüpft an die Debatte um die Demografisierung an, die sich damit beschäftigt, dass gesellschaftliche Entwicklungen häufig zu sehr aus einer demografischen Perspektive betrachtet und beschrieben werden. Sie analysiert die zum Teil noch jungen Diskurse, die durch Akteure in Wissenschaft, Politik und Publizistik zu den demografiepolitischen Strategien und Aktivitäten der Bundesregierung geführt werden.
Dabei werden Teilbereiche der Gesellschaft, darunter insbesondere die soziale Sicherung, die Generationenbeziehungen sowie die ökonomische Entwicklung und auch räumliche Aspekte der Demografie, in den Blick genommen. Es werden ausgewählte Zusammenhänge zwischen gesellschaftlichen Entwicklungen und demografischen Veränderungen reflektiert, denen ein kausales Beziehungsverhältnis bescheinigt wird. Es wird aufgezeigt, wo mit Verweis auf die Demografie lediglich Deutungsangebote geschaffen und Kausalbehauptungen aufgestellt werden.
Von besonderem Untersuchungsinteresse ist hierbei die Demografie als Argument, um politisches, wirtschaftliches und soziales Handeln zu legitimieren und ein gesellschaftliches Klima der Akzeptanz zu erzeugen. Wo erweist sich die Demografie als ein Spekulationsobjekt – wo ist sie erwiesene, überprüfbare Kausalität? Und wo ist die Grenze zur Instrumentalisierung der Demografie zu ziehen? Es wird belegt, dass eine Gelegenheitsdemografie vor allem aus drei Gründen praktiziert wird: Sie verschafft organisierten Interessen Gehör, sie bietet Orientierung in komplexen Gesellschaften und sie dient als Beurteilungsmaßstab der Bewertung von gesellschaftlichen Entwicklungen.
Die aktuelle Konjunktur von Diskursen zum demografischen Wandel sorgt dafür, dass die Gelegenheiten, mit Demografie zu argumentieren, immer wieder reproduziert werden. In der Folge werden nicht nur gesellschaftliche Entwicklungen zu sehr auf demografische Komponenten zurückgeführt. Auch das familien-, sozial-, migrations- und wirtschaftspolitische Denken und Handeln wird häufig über das Maß tatsächlicher Ursache-Wirkungs-Zusammenhänge hinaus auf vermeintliche demografische Gesetzmäßigkeiten reduziert (Reduktionismus).
Die Diskursanalyse der Demografiepolitik in Deutschland will die Bedeutung des demografischen Wandels für die Gesellschaft dabei nicht in Frage stellen. Sie soll für einen kritischeren Umgang mit der Demografie sensibilisieren. Dazu gehört auch, aufzuzeigen, dass die Demografie ein Faktor unter vielen ist.
This thesis investigates the application of polyelectrolyte multilayers in plasmonics and picosecond acoustics. The observed samples were fabricated by the spin-assisted layer-by-layer deposition technique that allowed a precise tuning of layer thickness in the range of few nanometers.
The first field of interest deals with the interaction of light-induced localized surface plasmons (LSP) of rod-shaped gold nanoparticles with the particles' environment. The environment consists of an air phase and a phase of polyelectrolytes, whose ratio affects the spectral position of the LSP resonance.
Measured UV-VIS spectra showed the shift of the LSP absorption peak as a function of the cover layer thickness of the particles. The data are modeled using an average dielectric function instead of the dielectric functions of air and polyelectrolytes. In addition using a measured dielectric function of the gold nanoparticles, the position of the LSP absorption peak could be simulated with good agreement to the data.
The analytic model helps to understand the optical properties of metal nanoparticles in an inhomogeneous environment.
The second part of this work discusses the applicability of PAzo/PAH and dye-doped PSS/PAH polyelectrolyte multilayers as transducers to generate hypersound pulses. The generated strain pulses were detected by time-domain Brillouin scattering (TDBS) using a pump-probe laser setup. Transducer layers made of polyelectrolytes were compared qualitatively to common aluminum transducers in terms of measured TDBS signal amplitude, degradation due to laser excitation, and sample preparation.
The measurements proved that fast and easy prepared polyelectrolyte transducers provided stronger TDBS signals than the aluminum transducer. AFM topography measurements showed a degradation of the polyelectrolyte structures, especially for the PAzo/PAH sample.
To quantify the induced strain, optical barriers were introduced to separate the transducer material from the medium of the hypersound propagation. Difficulties in the sample preparation prohibited a reliable quantification. But the experiments showed that a coating with transparent polyelectrolytes increases the efficiency of aluminum transducers and modifies the excited phonon distribution.
The adoption of polyelectrolytes to the scientific field of picosecond acoustics enables a cheap and fast fabrication of transducer layers on most surfaces. In contrast to aluminum layers the polyelectrolytes are transparent over a wide spectral range. Thus, the strain modulation can be probed from surface and back.
Das Schulbuch ist ein etablierter und bedeutender Bestandteil des Mathematikunterrichts. Lehrer nutzen es, um ihren Unterricht vorzubereiten und/oder zu gestalten; Schüler, um in selbigem zu lernen und zu bestehen, vielleicht sogar aus eigenem Interesse; Eltern, um sich darüber zu informieren, was ihr Kind eigentlich können soll und wie sie ihm gegebenenfalls helfen können. Darüber hinaus ist das Schulbuch ein markantes gesellschaftliches Produkt, dessen Zweck es ist, das Unterrichtsgeschehen zu steuern und zu beeinflussen. Damit ist es auch ein Anzeiger dafür, was und wie im Mathematikunterricht gelehrt werden sollte und wird. Die Lehrtexte als zentrale Bestandteile von Schulbüchern verweisen in diesem Zusammenhang insbesondere auf die Phasen der Einführung neuen Lernstoffs. Daraus legitimiert sich übergreifend die Fragestellung, was und wie (gut) Mathematikschulbuchlehrtexte lehren bzw. was und wie (gut) adressierte Schüler aus ihnen (selbstständig) lernen, d.h. Wissen erwerben können.
Angesichts der komplexen und vielfältigen Bedeutung von Schulbuchlehrtexten verwundert es, dass die mathematikdidaktische Forschung bislang wenig Interesse an ihnen zeigt: Es fehlen sowohl eine theoretische Konzeption der Größe ‚Lehrpotential eines schulmathematischen Lehrtextes‘ als auch ein analytisches Verfahren, um das anhand eines Mathematikschulbuchlehrtextes Verstehbare und Lernbare zu ermitteln. Mit der vorliegenden Arbeit wird sowohl in theoretisch-methodologischer als auch in empirischer Hinsicht der Versuch unternommen, diesen Defiziten zu begegnen. Dabei wird das ‚Lehrpotential eines Mathematikschulbuchlehrtextes‘ auf der Grundlage der kognitionspsychologischen Schematheorie und unter Einbeziehung textlinguistischer Ansätze als eine textimmanente und analytisch zugängliche Größe konzipiert. Anschließend wird das Lehrpotential von fünf Lehrtexten ausgewählter aktueller Schulbücher der Jahrgangsstufen 6 und 7 zu den Inhaltsbereichen ‚Brüche‘ und ‚lineare Funktionen‘ analysiert. Es zeigt sich, dass die untersuchten Lehrtexte aus deutschen Schulbüchern für Schüler sehr schwer verständlich sind, d.h. es ist kompliziert, einigen Teiltexten im Rahmen des Gesamttextes einen Sinn abzugewinnen. Die Lehrtexte sind insbesondere dann kaum sinnhaft lesbar, wenn ein Schüler versucht, die mitgeteilten Sachverhalte zu verstehen, d.h. Antworten auf die Fragen zu erhalten, warum ein mathematischer Sachverhalt gerade so und nicht anders ist, wozu ein neuer Sachverhalt/Begriff gebraucht wird, wie das Neue mit bereits Bekanntem zusammenhängt usw. Deutlich zugänglicher und sinnhafter erscheinen die Mathematikschulbuchlehrtexte hingegen unter der Annahme, dass ihre zentrale Botschaft in der Mitteilung besteht, welche Aufgabenstellungen in der jeweiligen Lehreinheit vorkommen und wie man sie bearbeitet. Demnach können Schüler anhand dieser Lehrtexte im Wesentlichen lernen, wie sie mit mathematischen Zeichen, die für sie kaum etwas bezeichnen, umgehen sollen. Die hier vorgelegten Analyseergebnisse gewinnen in einem soziologischen Kontext an Tragweite und Brisanz. So lässt sich aus ihnen u.a. die These ableiten, dass die analysierten Lehrtexte keine ‚unglücklichen‘ Einzelfälle sind, sondern dass die ‚Aufgabenorientierung in einem mathematischen Gewand‘ ein Charakteristikum typischer (deutscher) Mathematikschulbuchlehrtexte und – noch grundsätzlicher – einen Wesenszug typischer schulmathematischer Kommunikation darstellt.
Aufgrund ihrer potenziell gesundheitsfördernden Wirkung sind die polyphenolischen Isoflavone für die menschliche Ernährung von großem Interesse. Eine Vielzahl an experimentellen und epidemiologischen Studien zeigen für die in Soja enthaltenen Isoflavone Daidzein und Genistein eine präventive Wirkung bezüglich hormon-abhängiger und altersbedingter Erkrankungen, wie Brust- und Prostatakrebs, Osteoporose, Herz-Kreislauf-Erkrankungen sowie des menopausalen Syndroms. Die Metabolisierung und Bioaktivierung dieser sekundären Pflanzenstoffe durch die humane intestinale Darmmikrobiota ist individuell unterschiedlich. Nur in einem geringen Teil der westlichen Bevölkerung wird der Daidzein-Metabolit Equol durch spezifische Darmbakterien gebildet. Ein isoliertes Equol-produzierendes Bakterium des menschlichen Darmtrakts ist Slackia isoflavoniconvertens. Anhand dieser Spezies sollten die bislang unbekannten, an der Umsetzung von Daidzein und Genistein beteiligten Enzyme identifiziert und charakterisiert werden.
Fermentationsexperimente mit S. isoflavoniconvertens zeigten, dass die Gene der Daidzein und Genistein-umsetzenden Enzyme nicht konstitutiv exprimiert werden, sondern induziert werden müssen. Mit Hilfe der zweidimensionalen differentiellen Gelelektrophorese wurden sechs Proteine detektiert, welche in einer S. isoflavoniconvertens-Kultur in Anwesenheit von Daidzein induziert wurden. Auf Grundlage einzelner Peptidsequenzen erfolgte die Sequenzierung eines Genkomplexes mit den in gleicher Orientierung angeordneten Genen der durch Daidzein induzierten Proteine. Sequenzvergleiche identifizierten zudem äquivalente Genprodukte zu den Proteinen von S. isoflavoniconvertens in anderen Equolproduzierenden Bakterien. Nach der heterologen Expression in Escherichia coli wurden drei dieser Gene durch enzymatische Aktivitätstests als Daidzein-Reduktase (DZNR), Dihydrodaidzein-Reduktase (DHDR) und Tetrahydrodaidzein-Reduktase (THDR) identifiziert. Die Kombination der E. coli-Zellextrakte führte zur vollständigen Umsetzung von Daidzein über Dihydrodaidzein zu Equol. Neben Daidzein setzte die DZNR auch Genistein zu Dihydrogenistein um. Dies erfolgte mit einer größeren Umsatzgeschwindigkeit im Vergleich zur Reduktion von Daidzein zu Dihydrodaidzein. Enzymatische Aktivitätstests mit dem Zellextrakt von S. isoflavoniconvertens zeigten ebenfalls eine schnellere Umsetzung von Genistein. Die Kombination der rekombinanten DHDR und THDR führte zur Umsetzung von Dihydrodaidzein zu Equol. Der korrespondierende Metabolit 5-Hydroxyequol konnte als Endprodukt des Genistein-Metabolismus nicht detektiert werden. Zur Reinigung der drei identifizierten Reduktasen wurden diese genetisch an ein Strep-tag fusioniert und mittels Affinitätschromatographie gereinigt. Die übrigen durch Daidzein induzierten Proteine IfcA, IfcBC und IfcE wurden ebenfalls in E. coli exprimiert und als Strep-Fusionsproteine gereinigt. Vergleichende Aktivitätstests identifizierten das induzierte Protein IfcA als Dihydrodaidzein-Racemase. Diese katalysierte die Umsetzung des (R)- und (S)-Enantiomers von Dihydrodaidzein und Dihydrogenistein zum korrespondierenden Racemat. Neben dem Elektronentransfer-Flavoprotein IfcBC wurden auch die THDR, DZNR und IfcE als FAD-haltige Flavoproteine identifiziert. Zudem handelte es sich bei IfcE um ein Eisen-Schwefel-Protein. Nach Induktion der für die Daidzein-Umsetzung kodierenden Gene wurden mehrere verschieden lange mRNA-Transkripte gebildet. Dies zeigte, dass die Transkription des durch Daidzein induzierten Genkomplexes in S. isoflavoniconvertens nicht in Form eines einzelnen Operonsystems erfolgte.
Auf Grundlage der identifizierten Daidzein-umsetzenden Enzyme kann der Mechanismus der bakteriellen Umsetzung von Isoflavonen durch S. isoflavoniconvertens eingehend erforscht werden. Die ermittelten Gensequenzen der durch Daidzein induzierten Proteine sowie die korrespondierenden Gene weiterer Equol-produzierender Bakterien bieten zudem die Möglichkeit der mikrobiellen Metagenomanalyse im humanen Darmtrakt.
During the last two decades, instability training devices have become a popular means in athletic training and rehabilitation of mimicking unstable surfaces during movements like vertical jumps. Of note, under unstable conditions, trunk muscles seem to have a stabilizing function during exercise to facilitate the transfer of torques and angular momentum between the lower and upper extremities. The present thesis addresses the acute effects of surface instability on performance during jump-landing tasks. Additionally, the long-term effects (i.e., training) of surface instability were examined with a focus on the role of the trunk in athletic performance/physical fitness.
Healthy adolescent, and young adult subjects participated in three cross-sectional and one longitudinal study, respectively. Performance in jump-landing tasks on stable and unstable surfaces was assessed by means of a ground reaction force plate. Trunk muscle strength (TMS) was determined using an isokinetic device or the Bourban TMS test. Physical fitness was quantified by standing long jump, sprint, stand-and-reach, jumping sideways, Emery balance, and Y balance test on stable surfaces. In addition, activity of selected trunk and leg muscles and lower limb kinematics were recorded during jump-landing tasks.
When performing jump-landing tasks on unstable compared to stable surfaces, jump performance and leg muscle activity were significantly lower. Moreover, significantly smaller knee flexion angles and higher knee valgus angles were observed when jumping and landing on unstable compared to stable conditions and in women compared to men. Significant but small associations were found between behavioral and neuromuscular data, irrespective of surface condition. Core strength training on stable as well as on unstable surfaces significantly improved TMS, balance and coordination.
The findings of the present thesis imply that stable rather than unstable surfaces provide sufficient training stimuli during jump exercises (i.e., plyometrics). Additionally, knee motion strategy during plyometrics appears to be modified by surface instability and sex. Of note, irrespective of surface condition, trunk muscles only play a minor role for leg muscle performance/activity during jump exercises. Moreover, when implemented in strength training programs (i.e., core strength training), there is no advantage in using instability training devices compared to stable surfaces in terms of enhancement of athletic performance.
The main focus of the present thesis was to investigate the stabilization ability of poly(ionic liquid)s (PILs) in several examples as well as develop novel chemical structures and synthetic routes of PILs. The performed research can be specifically divided into three parts that include synthesis and application of hybrid material composed of PIL and cellulose nanofibers (CNFs), thiazolium-containing PILs, and main-chain imidazolium-type PILs.
In the first chapter, a vinylimidazolium-type IL was polymerized in water in the presence of CNFs resulting in the in situ electrostatic grafting of polymeric chains onto the surface of CNFs. The synthesized hybrid material merged advantages of its two components, that is, superior mechanical strength of CNFs and anion dependent solution properties of PILs. In contrast to unmodified CNFs, the hybrid could be stabilized and processed in organic solvents enabling its application as reinforcing agent for porous polyelectrolyte membranes.
In the second part, PILs and ionic polymers containing two types of thiazolium repeating units were synthesized. Such polymers displayed counterion dependent thermal stability and solubility in organic solvents of various dielectric constants. This new class of PILs was tested as stabilizers and phase transfer agents for carbon nanotubes in aqueous and organic media, and as binder materials to disperse electroactive powders and carbon additives in solid electrode in lithium-ion batteries. The incorporation of S and N atoms into the polymeric structures make such PILs also potential precursors for S, N - co-doped carbons.
In the last chapter, reactants originating from biomass were successfully harnessed to synthesize main-chain imidazolium-type PILs. An imidazolium-type diester IL obtained via a modified Debus-Radziszewski reaction underwent transesterification with diol in a polycondensation reaction. This yielded a polyester-type PIL which CO2 sorption properties were investigated. In the next step, the modified Debus-Radziszewski reaction was further applied to synthesize main-chain PILs according to a convenient, one-step protocol, using water as a green solvent and simple organic molecules as reagents. Depending on the structure of the employed diamine, the synthesized PILs after anion exchange showed superior thermal stability with unusually high carbonization yields.
Overall, the outcome of these studies will actively contribute to the current research on PILs by introducing novel PIL chemical structures, improved synthetic routes, and new examples of stabilized materials. The synthesis of main-chain imidazolium-type PILs by a modified Debus-Radziszewski reaction is of a special interest for the future work on porous ionic liquid networks as well as colloidal PIL nanoparticles.
Business Process Management has become an integral part of modern organizations in the private and public sector for improving their operations. In the course of Business Process Management efforts, companies and organizations assemble large process model repositories with many hundreds and thousands of business process models bearing a large amount of information. With the advent of large business process model collections, new challenges arise as structuring and managing a large amount of process models, their maintenance, and their quality assurance.
This is covered by business process architectures that have been introduced for organizing and structuring business process model collections. A variety of business process architecture approaches have been proposed that align business processes along aspects of interest, e. g., goals, functions, or objects. They provide a high level categorization of single processes ignoring their interdependencies, thus hiding valuable information. The production of goods or the delivery of services are often realized by a complex system of interdependent business processes. Hence, taking a holistic view at business processes interdependencies becomes a major necessity to organize, analyze, and assess the impact of their re-/design. Visualizing business processes interdependencies reveals hidden and implicit information from a process model collection.
In this thesis, we present a novel Business Process Architecture approach for representing and analyzing business process interdependencies on an abstract level. We propose a formal definition of our Business Process Architecture approach, design correctness criteria, and develop analysis techniques for assessing their quality. We describe a methodology for applying our Business Process Architecture approach top-down and bottom-up. This includes techniques for Business Process Architecture extraction from, and decomposition to process models while considering consistency issues between business process architecture and process model level. Using our extraction algorithm, we present a novel technique to identify and visualize data interdependencies in Business Process Data Architectures. Our Business Process Architecture approach provides business process experts,managers, and other users of a process model collection with an overview that allows reasoning about a large set of process models,
understanding, and analyzing their interdependencies in a facilitated way. In this regard we evaluated our Business Process Architecture approach in an experiment and provide implementations of selected techniques.
Was machen Schulleiter tatsächlich und welche Faktoren beeinflussen diese ausgeführten Tätigkeiten?
(2015)
Während die theoretische Arbeitsbeschreibung und das Rollenbild von Schulleitern vielfach in der Forschung aufgegriffen wurde, gibt es – wie übrigens im gesamten Bereich Public Management – nur wenige empirische Untersuchungen, die aus einer betriebswirtschaftlichen Managementbetrachtung heraus untersuchen, was Schulleiter wirklich machen, d.h. welchen Tätigkeiten und Aufgaben die genannten Personen nachgehen und welche Unterschiede sich feststellen lassen. Besondere Relevanz erhält die Thematik durch das sich wandelnde Aufgabenbild des Schulleiters, getrieben insbesondere durch die zusätzliche Autonomie der Einzelschule, aber auch durch die Fokussierung auf die Performance und Wirksamkeit der Einzelschule und verbunden damit, die Abhängigkeit dieser von der Arbeit des Schulleiters. Hier bildet das Verständnis der Aufgaben und Tätigkeiten eine wichtige Grundlage, die jedoch unzureichend erforscht ist. Mit Hilfe einer explorativen Beobachtung von 15 Schulleiterinnen und Schulleitern und damit einer empirischen Untersuchung von insgesamt 7591 Arbeitsminuten und 774 Aktivitäten in Kombination mit ausführlichen qualitativen, halboffenen Interviews wird durch diese Arbeit eine detaillierte Betrachtung des tatsächlichen Schulleitungsmanagementhandelns möglich. So wird sichtbar, dass sich die Aufgaben und Tätigkeiten der Schulleiter in zentralen Bereichen unterscheiden und eine Typologisierung entlang von Rollenbeschreibungen und Leadership Behavior zu kurz greift. Es konnte zum ersten Mal in dieser Ausführlichkeit innerhalb des deutschen Schulsystems gezeigt werden, dass Schulleiter Kommunikationsmanager sind. Darüber hinaus entwickelt das hier dokumentierte Forschungsvorhaben Hypothesen zu den Faktoren, die einen Einfluss auf die Aufgaben und Tätigkeiten haben und beschreibt dezidiert Implikationen, die diese Erkenntnisse auf die Tätigkeit des Schulleiters, die weitere Forschung aber auch die politische Rahmengestaltung und, damit verbunden, die Weiterentwicklung des Schulsystems haben.
Water resources from Central Asia’s mountain regions have a high relevance for the water supply of the water scarce lowlands. A good understanding of the water cycle in these mountain regions is therefore needed to develop water management strategies. Hydrological modeling helps to improve our knowledge of the regional water cycle, and it can be used to gain a better understanding of past changes or estimate future hydrologic changes in view of projected changes in climate. However, due to the scarcity of hydrometeorological data, hydrological modeling for mountain regions in Central Asia involves large uncertainties.
Addressing this problem, the first aim of this thesis was to develop hydrological modeling approaches that can increase the credibility of hydrological models in data sparse mountain regions. This was achieved by using additional data from remote sensing and atmospheric modeling. It was investigated whether spatial patterns from downscaled reanalysis data can be used for the interpolation of station-based precipitation data. This approach was compared to other precipitation estimates using a hydrologic evaluation based on hydrological modeling and a comparison of simulated and observed discharge, which demonstrated a generally good performance of this method. The study further investigated the value of satellite-derived snow cover data for model calibration. Trade-offs of good model performance in terms of discharge and snow cover were explicitly evaluated using a multiobjective optimization algorithm, and the results were contrasted with single-objective calibration and Monte Carlo simulations. The study clearly shows that the additional use of snow cover data improved the internal consistency of the hydrological model. In this context, it was further investigated for the first time how many snow cover scenes were required for hydrological model calibration.
The second aim of this thesis was the application of the hydrological model in order to investigate the causes of observed streamflow increases in two headwater catchments of the Tarim River over the recent decades. This simulation-based approach for trend attribution was complemented by a data-based approach. The hydrological model was calibrated to discharge and glacier mass balance data and considered changes in glacier geometry over time. The results show that in the catchment with a lower glacierization, increasing precipitation and temperature both contributed to the streamflow increases, while in the catchment with a stronger glacierization, increasing temperatures were identified as the dominant driver.
Most of the baryonic matter in the Universe resides in a diffuse gaseous phase in-between galaxies consisting mostly of hydrogen and helium. This intergalactic medium (IGM) is distributed in large-scale filaments as part of the overall cosmic web. The luminous extragalactic objects that we can observe today, such as galaxies and quasars, are surrounded by the IGM in the most dense regions within the cosmic web. The radiation of these objects contributes to the so-called ultraviolet background (UVB) which keeps the IGM highly ionized ever since the epoch of reionization.
Measuring the amount of absorption due to intergalactic neutral hydrogen (HI) against extragalactic background sources is a very useful tool to constrain the energy input of ionizing sources into the IGM. Observations suggest that the HI Lyman-alpha effective optical depth, τ_eff, decreases with decreasing redshift, which is primarily due to the expansion of the Universe. However, some studies find a smaller value of the effective optical depth than expected at the specific redshift z~3.2, possibly related to the complete reionization of helium in the IGM and a hardening of the UVB. The detection and possible cause of a decrease in τ_eff at z~3.2 is controversially debated in the literature and the observed features need further explanation.
To better understand the properties of the mean absorption at high redshift and to provide an answer for whether the detection of a τ_eff feature is real we study 13 high-resolution, high signal-to-noise ratio quasar spectra observed with the Ultraviolet and Visual Echelle Spectrograph (UVES) at the Very Large Telescope (VLT). The redshift evolution of the effective optical depth, τ_eff(z), is measured in the redshift range 2.7≤z≤3.6. The influence of metal absorption features is removed by performing a comprehensive absorption-line-fitting procedure.
In the first part of the thesis, a line-parameter analysis of the column density, N, and Doppler parameter, b, of ≈7500 individually fitted absorption lines is performed. The results are in good agreement with findings from previous surveys.
The second (main) part of this thesis deals with the analysis of the redshift evolution of the effective optical depth. The τ_eff measurements vary around the empirical power law τ_eff(z)~(1+z)^(γ+1) with γ=2.09±0.52. The same analysis as for the observed spectra is performed on synthetic absorption spectra. From a comparison between observed and synthetic spectral data it can be inferred that the uncertainties of the τ_eff values are likely underestimated and that the scatter is probably caused by high-column-density absorbers with column densities in the range 15≤logN≤17. In the real Universe, such absorbers are rarely observed, however. Hence, the difference in τ_eff from different observational data sets and absorption studies is most likely caused by cosmic variance. If, alternatively, the disagreement between such data is a result of an too optimistic estimate of the (systematic) errors, it is also possible that all τ_eff measurements agree with a smooth evolution within the investigated redshift range. To explore in detail the different analysis techniques of previous studies an extensive literature comparison to the results of this work is presented in this thesis.
Although a final explanation for the occurrence of the τ_eff deviation in different studies at z~3.2 cannot be given here, our study, which represents the most detailed line-fitting analysis of its kind performed at the investigated redshifts so far, represents another important benchmark for the characterization of the HI Ly-alpha effective optical depth at high redshift and its indicated unusual behavior at z~3.2.
Die Interaktionen von komplexen Kohlenhydraten und Proteinen sind ubiquitär. Sie spielen wichtige Rollen in vielen physiologischen Prozessen wie Zelladhäsion, Signaltransduktion sowie bei viralen Infektionen. Die molekularen Grundlagen der Interaktion sind noch nicht komplett verstanden. Ein Modellsystem für Kohlenhydrat-Protein-Interaktionen besteht aus Adhäsionsproteinen (Tailspikes) von Bakteriophagen, die komplexe Kohlenhydrate auf bakteriellen Oberflächen (O-Antigen) erkennen. Das Tailspike-Protein (TSP), das in dieser Arbeit betrachtet wurde, stammt aus dem Bakteriophagen 9NA (9NATSP). 9NATSP weist eine hohe strukturelle Homologie zum gut charakterisierten TSP des Phagen P22 (P22TSP) auf, bei einer niedriger sequenzieller Ähnlichkeit. Die Substratspezifitäten beider Tailspikes sind ähnlich mit Ausnahme der Toleranz gegenüber den glucosylierten Formen des O-Antigens. Die Struktur der beiden Tailspikes ist bekannt, sodass sie ein geeignetes System für vergleichende Bindungsstudien darstellen, um die strukturellen Grundlagen für die Unterschiede der Spezifität zu untersuchen.
Im Rahmen dieser Arbeit wurde der ELISA-like tailspike adsorption assay (ELITA) etabliert, um Binderpaare aus TSPs und O-Antigen zu identifizieren. Dabei wurden 9NATSP und P22TSP als Sonden eingesetzt, deren Bindung an die intakten, an die Mikrotiterplatte adsorbierten Bakterien getestet wurde. Beim Test einer Sammlung aus 44 Salmonella-Stämmen wurden Stämme identifiziert, die bindendes O-Antigen exprimieren. Gleichzeitig wurden Unterschiede in der Bindung der beiden TSPs an Salmonella-Stämme mit gleichem O-Serotyp beobachtet. Die Ergebnisse der ELITA-Messung wurden qualitativ durch eine FACS-basierte Bindungsmessung bestätigt. Zusätzlich ermöglichte die FACS-Messung bei Stämmen, die teilweise modifizierte O-Antigene herstellen, den Anteil an Zellen mit und ohne Modifikation zu erfassen.
Die Oberflächenplasmonresonanz (SPR)-basierten Interaktionsmessungen wurden eingesetzt, um Bindungsaffinitäten für eine TSP-O-Antigen Kombination zu quantifizieren. Dafür wurden zwei Methoden getestet, um die Oligosaccharide auf einem SPR-Chip zu immobilisieren. Zum einen wurden die enzymatisch hergestellten O-Antigenfragmente mit einem bifunktionalen Oxaminadapter derivatisiert, der eine primäre Aminogruppe für die Immobilisierung bereitstellt. Ein Versuch, diese Oligosaccharidfragmente zu immobilisieren, war jedoch nicht erfolgreich. Dagegen wurde das nicht derivatisierte Polysaccharid, bestehend aus repetitivem O-Antigen und einem konservierten Kernsaccharid, erfolgreich auf einem SPR-Chip immobilisiert. Die Immobilisierung wurde durch Interaktionsmessungen mit P22TSP bestätigt. Durch die Immobilisierung des Polysaccharids sind somit quantitative SPR-Bindungsmessungen mit einem polydispersen Interaktionspartner möglich.
Eine Auswahl von Salmonella-Stämmen mit einer ausgeprägt unterschiedlichen Bindung von 9NATSP und P22TSP im ELITA-Testsystem wurde hinsichtlich der Zusammensetzung des O-Antigens mittels HPLC, Kapillargelelektrophorese und MALDI-MS analysiert. Dabei wurden nicht-stöchiometrische Modifikationen der O-Antigene wie Acetylierung und Glucosylierung detektiert. Das Ausmaß der Glucosylierung korrelierte negativ mit der Effizienz der Bindung und des Verdaus durch die beiden TSPs, wobei der negative Effekt bei 9NATSP weniger stark ausgeprägt war als bei P22TSP. Dies stimmt mit den Literaturdaten zu Infektivitätsstudien mit 9NA und P22 überein, die mit Stämmen mit vergleichbaren O-Antigenvarianten durchgeführt wurden. Die Korrelation zwischen der Glucosylierung und Bindungseffizienz konnte strukturell interpretiert werden.
Auf Grundlage der O-Antigenanalysen sowie der Ergebnisse der ELITA- und FACS-Bindungstests wurden die Salmonella-Stämme Brancaster und Kalamu identifiziert, die annähernd quantitativ glucosyliertes O-Antigen exprimieren. Damit eignen sich diese Stämme für weiterführende Studien, um die Zusammenhänge zwischen der Spezifität und der Organisation der Bindestellen der beiden TSPs zu untersuchen.
Die Arbeit geht der Frage nach, wie Innovationen in einer Organisation des öffentlichen Sektors aufgenommen wurden und zu welchen Veränderungen dies führte. Im Vordergrund steht hier nicht die Innovation selbst, sondern vielmehr die Anpassungsmechanismen in der Organisation. Folgende Forschungsfragen wurden dazu gewählt:
1. Wie wurde das Instrument Zielsteuerung bzw. Zielvereinbarung im öffentlichen Sektor eingeführt und in die Managementroutinen integriert?
2. Welche Faktoren führen zu einer Integration der Zielsteuerung in die Managementroutinen?
3. Welche Empfehlungen für die Praxis lassen sich daraus ableiten?
Dazu wurde ein Landesbetrieb in Brandenburg detailliert untersucht und 31 Interviews mit Führungskräften der zweiten und dritten Managementebene geführt. In dieser Organisation wurde im Rahmen der deutschlandweiten Reformbewegung in der öffentlichen Verwaltung das Instrument Zielsteuerung bzw. Zielvereinbarung eingeführt und mit ganz konkreten Erwartungen verbunden. Als Untersuchungseinheit der möglichen Anpassungen und Veränderungen wurde das Konstrukt der Managementroutinen herangezogen, welche als kollektive Handlungsmuster ganz bewusst individuelle Verhaltensweisen ausklammerten.
Die Arbeit konnte eine Reihe von früheren Erkenntnissen bestätigen und zudem nachweisen, dass, entgegen des häufigen Vorurteils, Innovationen aus dem privatwirtschaftlichem Raum doch auch zu positiven Veränderungen in Organisationen der öffentlichen Hand führen können. Es kam hier jedoch nicht zur Entwicklung neuer, sondern zu einer Anpassung der bestehenden Routinen. Auf dieser Basis konnte festgestellt werden, dass ein stufenweiser Einführungsvorgang zunächst auf der Ebene der veränderten Zielvorstellungen der Führungskräfte zum Erfolg führte. Erst nach der Anpassung auf dieser „ostentativen“ Ebene kam es mit etwas Verzögerung zu einer Veränderung auf der Ebene der konkreten Handlungen. Im Hinblick auf die Einflussfaktoren der Innovation konnte festgestellt werden, dass viele Aspekte der Zielsetzungstheorie nach wie vor relevant sind und instabile politische Rahmenbedingungen zu wesentlichen Einschränkungen der Entfaltungsmöglichkeiten der Innovation führen können. Für viele Einflussfaktoren konnten allerdings sowohl positive als auch negative Wirkungen identifiziert werden.
Aminosäuren sind lebensnotwendige Moleküle für alle Organismen. Ihre Erkennung im Körper ermöglicht eine bedarfsgerechte Regulation ihrer Aufnahme und ihrer Verwertung. Welcher Chemosensor für diese Erkennung jedoch hauptverantwortlich ist, ist bisher unklar. In der vorliegenden Arbeit wurde die Rolle der Umamigeschmacksrezeptoruntereinheit Tas1r1 jenseits ihrer gustatorischen Bedeutung für die Aminosäuredetektion in der Mundhöhle untersucht.
In der histologischen Tas1r1-Expressionsanalyse nichtgustatorischer Gewebe der Mauslinie Tas1r1-Cre/ROSA26-tdRFP wurde über die Detektion des Reporterproteins tdRFP die Expression des Tas1r1 in allen untersuchten Geweben (Speiseröhre, Magen, Darm, Bauchspeicheldrüse, Leber, Niere, Muskel- und Fettgewebe, Milz, Thymus, Lymphknoten, Lunge sowie Hoden) nachgewiesen. Mit Ausnahme von Dünndarm und Hoden gelang hierbei der Nachweis erstmals spezifisch auf zellulärer Ebene. Caecum und Lymphknoten wurden zudem neu als Expressionsorte des Tas1r1 identifiziert.
Trotz der beobachteten weiten Verbreitung des Tas1r1 im Organismus – unter anderem auch in Geweben, die für den Proteinstoffwechsel besonders relevant sind – waren im Zuge der durchgeführten Untersuchung potentieller extraoraler Funktionen des Rezeptors durch phänotypische Charakterisierung der Mauslinie Tas1r1-BLiR nur schwache Auswirkungen auf Aminosäurestoffwechsel bzw. Stickstoffhaushalt im Falle eines Tas1r1-Knockouts detektierbar. Während sich Ernährungsverhalten, Gesamtphysiologie, Gewebemorphologie sowie Futterverdaulichkeit unverändert zeigten, war die renale Stickstoffausscheidung bei Tas1r1-Knockout-Mäusen auf eiweißarmer sowie auf eiweißreicher Diät signifikant verringert. Eine Überdeckung der Auswirkungen des Tas1r1-Knockouts aufgrund kompensatorischer Effekte durch den Aminosäuresensor CaSR oder den Peptidsensor Gpr93 war nicht nachweisbar. Es bleibt offen, ob andere Mechanismen oder andere Chemosensoren an einer Kompensation beteiligt sind oder aber Tas1r1 in extraoralem Gewebe andere Funktionen als die der Aminosäuredetektion übernimmt. Unterschiede im extraoralen Expressionsmuster der beiden Umamirezeptor-untereinheiten Tas1r1 und Tasr3 lassen Spekulationen über andere Partner, Liganden und Funktionen zu.