Refine
Has Fulltext
- yes (161) (remove)
Year of publication
- 2011 (161) (remove)
Document Type
- Doctoral Thesis (73)
- Postprint (40)
- Article (18)
- Monograph/Edited Volume (12)
- Part of a Book (6)
- Master's Thesis (3)
- Working Paper (3)
- Habilitation Thesis (2)
- Preprint (2)
- Other (1)
Language
- English (161) (remove)
Keywords
- Holocene (3)
- Holozän (3)
- Klimawandel (3)
- Modellierung (3)
- Selbstorganisation (3)
- answer set programming (3)
- climate change (3)
- eye movements (3)
- self-assembly (3)
- Blickbewegungen (2)
Institute
- Institut für Biochemie und Biologie (19)
- Institut für Physik und Astronomie (18)
- Mathematisch-Naturwissenschaftliche Fakultät (18)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (15)
- Extern (14)
- Institut für Geowissenschaften (14)
- Institut für Chemie (13)
- Wirtschaftswissenschaften (11)
- Institut für Informatik und Computational Science (8)
- Sonderforschungsbereich 632 - Informationsstruktur (6)
- Department Linguistik (5)
- Department Psychologie (5)
- Institut für Anglistik und Amerikanistik (5)
- Institut für Künste und Medien (5)
- Institut für Jüdische Studien und Religionswissenschaft (4)
- Vereinigung für Jüdische Studien e. V. (4)
- Humanwissenschaftliche Fakultät (3)
- Institut für Mathematik (3)
- Institut für Umweltwissenschaften und Geographie (2)
- Philosophische Fakultät (2)
- Sozialwissenschaften (2)
- Institut für Ernährungswissenschaft (1)
- Institut für Germanistik (1)
- Institut für Philosophie (1)
- Institut für Romanistik (1)
- Juristische Fakultät (1)
- MenschenRechtsZentrum (1)
- Strukturbereich Kognitionswissenschaften (1)
- WeltTrends e.V. Potsdam (1)
- Wirtschafts- und Sozialwissenschaftliche Fakultät (1)
Define real, Moron!
(2011)
Academic language should not be a ghetto dialect at odds with ordinary language, but rather an extension that is compatible with lay-language. To define ‘game’ with the unrealistic ambition of satisfying both lay-people and experts should not be a major concern for a game ontology, since the field it addresses is subject to cultural evolution and diachronic change. Instead of the impossible mission of turning the common word into an analytic concept, a useful task for an ontology of games is to model game differences, to show how the things we call games can be different from each other in a number of different ways.
Unique column combinations of a relational database table are sets of columns that contain only unique values. Discovering such combinations is a fundamental research problem and has many different data management and knowledge discovery applications. Existing discovery algorithms are either brute force or have a high memory load and can thus be applied only to small datasets or samples. In this paper, the wellknown GORDIAN algorithm and "Apriori-based" algorithms are compared and analyzed for further optimization. We greatly improve the Apriori algorithms through efficient candidate generation and statistics-based pruning methods. A hybrid solution HCAGORDIAN combines the advantages of GORDIAN and our new algorithm HCA, and it significantly outperforms all previous work in many situations.
Yoel Hoffmann is an Israeli writer born in 1937 in Brasov (Kronstadt), Romania. Brought up in a German-speaking family, already in his first book, Sefer Yosef (1989), he conveys the voice of German-speaking immigrants in Israel (the “Katschen” story, 1986) and that of the East European Jewish community in Berlin in the late 1930s, on the verge of the Second World War. His works are crammed with characters of Jews from Germany gripped by the memory of the language they abandoned following their emigration to Palestine in the 1930s. The classic one is the character of Bernhard, in the eponymous work. The current article focuses on the representation and elaboration of Hoffmann’s unique creation, in a language influenced by his deep identification with Zen Buddhism on the one hand, and his attraction to the modernist, Western style of stream of consciousness on the other. In central sections of his works, Hoffman presents his entire literary corpus as a type of explicit, allusive, or secret Holocaust literature, and invites his readers and his critics to decode the allusions and expose the secret in this theme, a surprising statement in relation to Hoffmann’s work and its analysis so far. Hoffmann represents the Holocaust as a collective Israeli trauma for which his literary fiction creates a special catalogue of representative characters. In the creation of a catalogue, and particularly one that simultaneously classifies and individualizes, Hoffmann’s project resembles the monumental 1920s cataloguing project by the celebrated German photographer August Sander (Menschen des 20. Jahrhunderts). Hoffmann included photographs from this project in his works, and even chose some of them for the covers of his books. The article examines the implicit relationships between these two creative artists as conferring a meaning so far not considered in the research of the Holocaust theme in Yoel Hoffmann’s writings.
Background
Riociguat is the first of a new class of drugs, the soluble guanylate cyclase (sGC) stimulators. Riociguat has a dual mode of action: it sensitizes sGC to the body’s own NO and can also increase sGC activity in the absence of NO. The NO-sGC-pathway is impaired in many cardiovascular diseases such as heart failure, pulmonary hypertension and diabetic nephropathy (DN). DN leads to high cardiovascular morbidity and mortality. There is still a high unmet medical need. The urinary albumin excretion rate is a predictive biomarker for these clinical events. Therefore, we investigated the effect of riociguat, alone and in combination with the angiotensin II receptor antagonist (ARB) telmisartan on the progression of DN in diabetic eNOS knock out mice, a new model closely resembling human pathology.
Methods
Seventy-six male eNOS knockout C57BL/6J mice were divided into 4 groups after receiving intraperitoneal high-dose streptozotocin: telmisartan (1 mg/kg), riociguat (3 mg/kg), riociguat+telmisartan (3 and 1 mg/kg), and vehicle. Fourteen mice were used as non-diabetic controls. After 12 weeks, urine and blood were obtained and blood pressure measured. Glucose concentrations were highly increased and similar in all diabetic groups.
Results
Riociguat, alone (105.2 ± 2.5 mmHg; mean±SEM; n = 14) and in combination with telmisartan (105.0 ± 3.2 mmHg; n = 12), significantly reduced blood pressure versus diabetic controls (117.1 ± 2.2 mmHg; n = 14; p = 0.002 and p = 0.004, respectively), whereas telmisartan alone (111.2 ± 2.6 mmHg) showed a modest blood pressure lowering trend (p = 0.071; n = 14). The effects of single treatment with either riociguat (97.1 ± 15.7 µg/d; n = 13) or telmisartan (97.8 ± 26.4 µg/d; n = 14) did not significantly lower albumin excretion on its own (p = 0.067 and p = 0.101, respectively). However, the combined treatment led to significantly lower urinary albumin excretion (47.3 ± 9.6 µg/d; n = 12) compared to diabetic controls (170.8 ± 34.2 µg/d; n = 13; p = 0.004), and reached levels similar to non-diabetic controls (31.4 ± 10.1 µg/d, n = 12).
Conclusion
Riociguat significantly reduced urinary albumin excretion in diabetic eNOS knock out mice that were refractory to treatment with ARB’s alone. Patients with diabetic nephropathy refractory to treatment with ARB’s have the worst prognosis among all patients with diabetic nephropathy. Our data indicate that additional stimulation of sGC on top of standard treatment with ARB`s may offer a new therapeutic approach for patients with diabetic nephropathy resistant to ARB treatment.
A systems biological approach towards the molecular basis of heterosis in Arabidopsis thaliana
(2011)
Heterosis is defined as the superiority in performance of heterozygous genotypes compared to their corresponding genetically different homozygous parents. This phenomenon is already known since the beginning of the last century and it has been widely used in plant breeding, but the underlying genetic and molecular mechanisms are not well understood. In this work, a systems biological approach based on molecular network structures is proposed to contribute to the understanding of heterosis. Hybrids are likely to contain additional regulatory possibilities compared to their homozygous parents and, therefore, they may be able to correctly respond to a higher number of environmental challenges, which leads to a higher adaptability and, thus, the heterosis phenomenon. In the network hypothesis for heterosis, presented in this work, more regulatory interactions are expected in the molecular networks of the hybrids compared to the homozygous parents. Partial correlations were used to assess this difference in the global interaction structure of regulatory networks between the hybrids and the homozygous genotypes. This network hypothesis for heterosis was tested on metabolite profiles as well as gene expression data of the two parental Arabidopsis thaliana accessions C24 and Col-0 and their reciprocal crosses. These plants are known to show a heterosis effect in their biomass phenotype. The hypothesis was confirmed for mid-parent and best-parent heterosis for either hybrid of our experimental metabolite as well as gene expression data. It was shown that this result is influenced by the used cutoffs during the analyses. Too strict filtering resulted in sets of metabolites and genes for which the network hypothesis for heterosis does not hold true for either hybrid regarding mid-parent as well as best-parent heterosis. In an over-representation analysis, the genes that show the largest heterosis effects according to our network hypothesis were compared to genes of heterotic quantitative trait loci (QTL) regions. Separately for either hybrid regarding mid-parent as well as best-parent heterosis, a significantly larger overlap between the resulting gene lists of the two different approaches towards biomass heterosis was detected than expected by chance. This suggests that each heterotic QTL region contains many genes influencing biomass heterosis in the early development of Arabidopsis thaliana. Furthermore, this integrative analysis led to a confinement and an increased confidence in the group of candidate genes for biomass heterosis in Arabidopsis thaliana identified by both approaches.
Non-mycorrhizal fungal endophytes are able to colonize internally roots without causing visible disease symptoms establishing neutral or mutualistic associations with plants. These fungi known as non-clavicipitaceous endophytes have a broad host range of monocot and eudicot plants and are highly diverse. Some of them promote plant growth and confer increased abiotic-stress tolerance and disease resistance. According to such possible effects on host plants, it was aimed to isolate and to characterize native fungal root endophytes from tomato (Lycopersicon esculentum Mill.) and to analyze their effects on plant development, plant resistance and fruit yield and quality together with the model endophyte Piriformospora indica. Fifty one new fungal strains were isolated from desinfected tomato roots of four different crop sites in Colombia. These isolates were roughly characterized and fourteen potential endophytes were further analyzed concerning their taxonomy, their root colonization capacity and their impact on plant growth. Sequencing of the ITS region from the ribosomal RNA gene cluster and in-depth morphological characterisation revealed that they correspond to different phylogenetic groups among the phylum Ascomycota. Nine different morphotypes were described including six dark septate endophytes (DSE) that did not correspond to the Phialocephala group. Detailed confocal microscopy analysis showed various colonization patterns of the endophytes inside the roots ranging from epidermal penetration to hyphal growth through the cortex. Tomato pot experiments under glass house conditions showed that they differentially affect plant growth depending on colonization time and inoculum concentration. Three new isolates (two unknown fungal endophyte DSE48, DSE49 and one identified as Leptodontidium orchidicola) with neutral or positiv effects were selected and tested in several experiments for their influence on vegetative growth, fruit yield and quality and their ability to diminish the impact of the pathogen Verticillium dahliae on tomato plants. Although plant growth promotion by all three fungi was observed in young plants, vegetative growth parameters were not affected after 22 weeks of cultivation except a reproducible increase of root diameter by the endophyte DSE49. Additionally, L. orchidicola increased biomass and glucose content of tomato fruits, but only at an early date of harvest and at a certain level of root colonization. Concerning bioprotective effects, the endophytes DSE49 and L. orchidicola decreased significantly disease symptoms caused by the pathogen V. dahliae, but only at a low dosis of the pathogen. In order to analyze, if the model root endophytic fungus Piriformospora indica could be suitable for application in production systems, its impact on tomato was evaluated. Similarly to the new fungal isolates, significant differences for vegetative growth parameters were only observable in young plants and, but protection against V. dahliae could be seen in one experiment also at high dosage of the pathogen. As the DSE L. orchidicola, P. indica increased the number and biomass of marketable tomatoes only at the beginning of fruit setting, but this did not lead to a significant higher total yield. If the effects on growth are due to a better nutrition of the plant with mineral element was analyzed in barley in comparison to the arbuscular mycorrhizal fungus Glomus mosseae. While the mycorrhizal fungus increased nitrogen and phosphate uptake of the plant, no such effect was observed for P. indica. In summary this work shows that many different fungal endophytes can be also isolated from roots of crops and, that these isolates can have positive effects on early plant development. This does, however, not lead to an increase in total yield or in improvement of fruit quality of tomatoes under greenhouse conditions.
This paper develops the incentives to collude in a model with spatially separated markets and quantity setting firms. We find that increases in transportation costs stabilize the collusive agreement. We also show that, the higher the demand in both markets the less likely will collusion be sustained. Gross and Holahan (2003) use a similar model with price setting firms, we compare their results with ours to analyze the impact of the mode of competition on sustainability of collusion. Further we analyze the impact of collusion on social welfare and find that collusion may be welfare enhancing.
Corvino, Corvino and Schoen, Chruściel and Delay have shown the existence of a large class of asymptotically flat vacuum initial data for Einstein's field equations which are static or stationary in a neighborhood of space-like infinity, yet quite general in the interior. The proof relies on some abstract, non-constructive arguments which makes it difficult to calculate such data numerically by using similar arguments. A quasilinear elliptic system of equations is presented of which we expect that it can be used to construct vacuum initial data which are asymptotically flat, time-reflection symmetric, and asymptotic to static data up to a prescribed order at space-like infinity. A perturbation argument is used to show the existence of solutions. It is valid when the order at which the solutions approach staticity is restricted to a certain range. Difficulties appear when trying to improve this result to show the existence of solutions that are asymptotically static at higher order. The problems arise from the lack of surjectivity of a certain operator. Some tensor decompositions in asymptotically flat manifolds exhibit some of the difficulties encountered above. The Helmholtz decomposition, which plays a role in the preparation of initial data for the Maxwell equations, is discussed as a model problem. A method to circumvent the difficulties that arise when fast decay rates are required is discussed. This is done in a way that opens the possibility to perform numerical computations. The insights from the analysis of the Helmholtz decomposition are applied to the York decomposition, which is related to that part of the quasilinear system which gives rise to the difficulties. For this decomposition analogous results are obtained. It turns out, however, that in this case the presence of symmetries of the underlying metric leads to certain complications. The question, whether the results obtained so far can be used again to show by a perturbation argument the existence of vacuum initial data which approach static solutions at infinity at any given order, thus remains open. The answer requires further analysis and perhaps new methods.
Immobile Tremor
(2011)
The threshold between the XVth and the XVIth Century represents a historical period during which, both for Christians and for Jews, the geopolitical sceneries and the interior horizons radically change. The modified reality provokes new forms of expectation and the need of new historical interpretations. Ferrara, within this scenery, can be considered, as other Italian cases, as a paradigmatic example, a narrow space where phenomena of spiritual and cultural Jewish rebirth can take shape. The permeability between Christian artistic and cultural world and Jewish intellectual production determines a prosperous context, further strengthened by the introduction of Jewish typography and by a growing claim and restoration of social elective dignity among the Jews of the Este Duchy. After the transfer of the capital city from Ferrara to Modena, the indirect effects of this intellectual resurgence are deeply transformed on a social level, and allows us to catch the persistence of important forms of communication between Christians and Jews in everyday life. The introduction of the Inquisition provides us, through the production of the judicial archive, with the most important instrument to understand social dynamics, which allows us to comprehend a new potential interpretation key for the reality of the ghetto and the choice of its erection. The urban division is nothing else but a new attempt to separate the invisible spaces of the thought. The effective efficacy of the physical separation shows several weak points, which persist during the entire life of the ghetto, since 1638 until 1789.
Actin-based directional motility is important for embryonic development, wound healing, immune responses, and development of tissues. Actin and myosin are essential players in this process that can be subdivided into protrusion, adhesion, and traction. Protrusion is the forward movement of the membrane at the leading edge of the cell. Adhesion is required to enable movement along a substrate, and traction finally leads to the forward movement of the entire cell body, including its organelles. While actin polymerization is the main driving force in cell protrusions, myosin motors lead to the contraction of the cell body. The goal of this work was to study the regulatory mechanisms of the motile machinery by selecting a representative key player for each stage of the signaling process: the regulation of Arp2/3 activity by WASP (actin system), the role of cGMP in myosin II assembly (myosin system), and the influence of phosphoinositide signaling (upstream receptor pathway). The model organism chosen for this work was the social ameba Dictyostelium discoideum, due to the well-established knowledge of its cytoskeletal machinery, the easy handling, and the high motility of its vegetative and starvation developed cells. First, I focused on the dynamics of the actin cytoskeleton by modulating the activity of one of its key players, the Arp2/3 complex. This was achieved using the carbazole derivative Wiskostatin, an inhibitor of the Arp2/3 activator WASP. Cells treated with Wiskostatin adopted a round shape, with no of few pseudopodia. With the help of a microfluidic cell squeezer device, I could show that Wiskostatin treated cells display a reduced mechanical stability, comparable to cells treated with the actin disrupting agent Latrunculin A. Furthermore, the WASP inhibited cells adhere stronger to a surface and show a reduced motility and chemotactic performance. However, the overall F-actin content in the cells was not changed. Confocal microscopy and TIRF microscopy imaging showed that the cells maintained an intact actin cortex. Localized dynamic patches of increased actin polymerization were observed that, however, did not lead to membrane deformation. This indicated that the mechanisms of actin-driven force generation were impaired in Wiskostatin treated cells. It is concluded that in these cells, an altered architecture of the cortical network leads to a reduced overall stiffness of the cell, which is insufficient to support the force generation required for membrane deformation and pseudopod formation. Second, the role of cGMP in myosin II dynamics was investigated. Cyclic GMP is known to regulate the association of myosin II with the cytoskeleton. In Dictyostelium, intracellular cGMP levels increase when cells are exposed to chemoattractants, but also in response to osmotic stress. To study the influence of cyclic GMP on actin and myosin II dynamics, I used the laser-induced photoactivation of a DMACM-caged-Br-cGMP to locally release cGMP inside the cell. My results show that cGMP directly activates the myosin II machinery, but is also able to induce an actin response independently of cAMP receptor activation and signaling. The actin response was observed in both vegetative and developed cells. Possible explanations include cGMP-induced actin polymerization through VASP (vasodilator-stimulated phosphoprotein) or through binding of cGMP to cyclic nucleotide-dependent kinases. Finally, I investigated the role of phosphoinositide signaling using the Polyphosphoinositide-Binding Peptide (PBP10) that binds preferentially to PIP2. Phosphoinositides can recruit actin-binding proteins to defined subcellular sites and alter their activity. Neutrophils, as well as developed Dictyostelium cells produce PIP3 in the plasma membrane at their leading edge in response to an external chemotactic gradient. Although not essential for chemotaxis, phosphoinositides are proposed to act as an internal compass in the cell. When treated with the peptide PBP10, cells became round, with fewer or no pseudopods. PH-CRAC translocation to the membrane still occurs, even at low cAMP stimuli, but cell motility (random and directional) was reduced. My data revealed that the decrease in the pool of available PIP2 in the cell is sufficient to impair cell motility, but enough PIP2 remains so that PIP3 is formed in response to chemoattractant stimuli. My data thus highlights how sensitive cell motility and morphology are to changes in the phosphoinositide signaling. In summary, I have analyzed representative regulatory mechanisms that govern key parts of the motile machinery and characterized their impact on cellular properties including mechanical stability, adhesion and chemotaxis.
In the present work synchronization phenomena in complex dynamical systems exhibiting multiple time scales have been analyzed. Multiple time scales can be active in different manners. Three different systems have been analyzed with different methods from data analysis. The first system studied is a large heterogenous network of bursting neurons, that is a system with two predominant time scales, the fast firing of action potentials (spikes) and the burst of repetitive spikes followed by a quiescent phase. This system has been integrated numerically and analyzed with methods based on recurrence in phase space. An interesting result are the different transitions to synchrony found in the two distinct time scales. Moreover, an anomalous synchronization effect can be observed in the fast time scale, i.e. there is range of the coupling strength where desynchronization occurs. The second system analyzed, numerically as well as experimentally, is a pair of coupled CO₂ lasers in a chaotic bursting regime. This system is interesting due to its similarity with epidemic models. We explain the bursts by different time scales generated from unstable periodic orbits embedded in the chaotic attractor and perform a synchronization analysis of these different orbits utilizing the continuous wavelet transform. We find a diverse route to synchrony of these different observed time scales. The last system studied is a small network motif of limit cycle oscillators. Precisely, we have studied a hub motif, which serves as elementary building block for scale-free networks, a type of network found in many real world applications. These hubs are of special importance for communication and information transfer in complex networks. Here, a detailed study on the mechanism of synchronization in oscillatory networks with a broad frequency distribution has been carried out. In particular, we find a remote synchronization of nodes in the network which are not directly coupled. We also explain the responsible mechanism and its limitations and constraints. Further we derive an analytic expression for it and show that information transmission in pure phase oscillators, such as the Kuramoto type, is limited. In addition to the numerical and analytic analysis an experiment consisting of electrical circuits has been designed. The obtained results confirm the former findings.
Background
In many species males face a higher predation risk than females because males display elaborate traits that evolved under sexual selection, which may attract not only females but also predators. Females are, therefore, predicted to avoid such conspicuous males under predation risk. The present study was designed to investigate predator-induced changes of female mating preferences in Atlantic mollies (Poecilia mexicana). Males of this species show a pronounced polymorphism in body size and coloration, and females prefer large, colorful males in the absence of predators.
Results
In dichotomous choice tests predator-naïve (lab-reared) females altered their initial preference for larger males in the presence of the cichlid Cichlasoma salvini, a natural predator of P. mexicana, and preferred small males instead. This effect was considerably weaker when females were confronted visually with the non-piscivorous cichlid Vieja bifasciata or the introduced non-piscivorous Nile tilapia (Oreochromis niloticus). In contrast, predator experienced (wild-caught) females did not respond to the same extent to the presence of a predator, most likely due to a learned ability to evaluate their predators' motivation to prey.
Conclusions
Our study highlights that (a) predatory fish can have a profound influence on the expression of mating preferences of their prey (thus potentially affecting the strength of sexual selection), and females may alter their mate choice behavior strategically to reduce their own exposure to predators. (b) Prey species can evolve visual predator recognition mechanisms and alter their mate choice only when a natural predator is present. (c) Finally, experiential effects can play an important role, and prey species may learn to evaluate the motivational state of their predators.
In the living cell, the organization of the complex internal structure relies to a large extent on molecular motors. Molecular motors are proteins that are able to convert chemical energy from the hydrolysis of adenosine triphosphate (ATP) into mechanical work. Being about 10 to 100 nanometers in size, the molecules act on a length scale, for which thermal collisions have a considerable impact onto their motion. In this way, they constitute paradigmatic examples of thermodynamic machines out of equilibrium. This study develops a theoretical description for the energy conversion by the molecular motor myosin V, using many different aspects of theoretical physics. Myosin V has been studied extensively in both bulk and single molecule experiments. Its stepping velocity has been characterized as a function of external control parameters such as nucleotide concentration and applied forces. In addition, numerous kinetic rates involved in the enzymatic reaction of the molecule have been determined. For forces that exceed the stall force of the motor, myosin V exhibits a 'ratcheting' behaviour: For loads in the direction of forward stepping, the velocity depends on the concentration of ATP, while for backward loads there is no such influence. Based on the chemical states of the motor, we construct a general network theory that incorporates experimental observations about the stepping behaviour of myosin V. The motor's motion is captured through the network description supplemented by a Markov process to describe the motor dynamics. This approach has the advantage of directly addressing the chemical kinetics of the molecule, and treating the mechanical and chemical processes on equal grounds. We utilize constraints arising from nonequilibrium thermodynamics to determine motor parameters and demonstrate that the motor behaviour is governed by several chemomechanical motor cycles. In addition, we investigate the functional dependence of stepping rates on force by deducing the motor's response to external loads via an appropriate Fokker-Planck equation. For substall forces, the dominant pathway of the motor network is profoundly different from the one for superstall forces, which leads to a stepping behaviour that is in agreement with the experimental observations. The extension of our analysis to Markov processes with absorbing boundaries allows for the calculation of the motor's dwell time distributions. These reveal aspects of the coordination of the motor's heads and contain direct information about the backsteps of the motor. Our theory provides a unified description for the myosin V motor as studied in single motor experiments.
The present thesis was born and evolved within the RAdial Velocity Experiment (RAVE) with the goal of measuring chemical abundances from the RAVE spectra and exploit them to investigate the chemical gradients along the plane of the Galaxy to provide constraints on possible Galactic formation scenarios. RAVE is a large spectroscopic survey which aims to observe spectroscopically ~10^6 stars by the end of 2012 and measures their radial velocities, atmospheric parameters and chemical abundances. The project makes use of the UK Schmidt telescope at Australian Astronomical Observatory (AAO) in Siding Spring, Australia, equipped with the multiobject spectrograph 6dF. To date, RAVE collected and measured more than 450,000 spectra. The precision of the chemical abundance estimations depends on the reliability of the atomic and atmosphere parameters adopted (in particular the oscillator strengths of the absorption lines and the effective temperature, gravity, and metallicity of the stars measured). Therefore we first identified 604 absorption lines in the RAVE wavelength range and refined their oscillator strengths with an inverse spectral analysis. Then, we improved the RAVE stellar parameters by modifying the RAVE pipeline and the spectral library the pipeline rely on. The modifications removed some systematic errors in stellar parameters discovered during this work. To obtain chemical abundances, we developed two different processing pipelines. Both of them perform chemical abundances measurements by assuming stellar atmospheres in Local Thermodynamic Equilibrium (LTE). The first one determines elements abundances from equivalent widths of absorption lines. Since this pipeline showed poor sensibility on abundances relative to iron, it has been superseded. The second one exploits the chi^2 minimization technique between observed and model spectra. Thanks to its precision, it has been adopted for the creation of the RAVE chemical catalogue. This pipeline provides abundances with uncertains of about ~0.2dex for spectra with signal-to-noise ratio S/N>40 and ~0.3dex for spectra with 20>S/N>40. For this work, the pipeline measured chemical abundances up to 7 elements for 217,358 RAVE stars. With these data we investigated the chemical gradients along the Galactic radius of the Milky Way. We found that stars with low vertical velocities |W| (which stay close to the Galactic plane) show an iron abundance gradient in agreement with previous works (~-0.07$ dex kpc^-1) whereas stars with larger |W| which are able to reach larger heights above the Galactic plane, show progressively flatter gradients. The gradients of the other elements follow the same trend. This suggests that an efficient radial mixing acts in the Galaxy or that the thick disk formed from homogeneous interstellar matter. In particular, we found hundreds of stars which can be kinetically classified as thick disk stars exhibiting a chemical composition typical of the thin disk. A few stars of this kind have already been detected by other authors, and their origin is still not clear. One possibility is that they are thin disk stars kinematically heated, and then underwent an efficient radial mixing process which blurred (and so flattened) the gradient. Alternatively they may be a transition population" which represents an evolutionary bridge between thin and thick disk. Our analysis shows that the two explanations are not mutually exclusive. Future follow-up high resolution spectroscopic observations will clarify their role in the Galactic disk evolution.
Parsability approaches of several grammar formalisms generating also non-context-free languages are explored. Chomsky grammars, Lindenmayer systems, grammars with controlled derivations, and grammar systems are treated. Formal properties of these mechanisms are investigated, when they are used as language acceptors. Furthermore, cooperating distributed grammar systems are restricted so that efficient deterministic parsing without backtracking becomes possible. For this class of grammar systems, the parsing algorithm is presented and the feature of leftmost derivations is investigated in detail.
Eye fixation durations during normal reading correlate with processing difficulty but the specific cognitive mechanisms reflected in these measures are not well understood. This study finds support in German readers’ eyefixations for two distinct difficulty metrics: surprisal, which reflects the change in probabilities across syntactic analyses as new words are integrated, and retrieval, which quantifies comprehension difficulty in terms of working memory constraints. We examine the predictions of both metrics using a family of dependency parsers indexed by an upper limit on the number of candidate syntactic analyses they retain at successive words. Surprisal models all fixation measures and regression probability. By contrast, retrieval does not model any measure in serial processing. As more candidate analyses are considered in parallel at each word, retrieval can account for the same measures as surprisal. This pattern suggests an important role for ranked parallelism in theories of sentence comprehension.
Dutch allows for variation as to whether the first position in the sentence is occupied by the subject or by some other constituent, such as the direct object. In particular situations, however, this commonly observed variation in word order is ‘frozen’ and only the subject appears in first position. We hypothesize that this partial freezing of word order in Dutch can be explained from the dependence of the speaker’s choice of word order on the hearer’s interpretation of this word order. A formal model of this interaction between the speaker’s perspective and the hearer’s perspective is presented in terms of bidirectional Optimality Theory. Empirical predictions of this model regarding the interaction between word order and definiteness are confirmed by a quantitative corpus study.
Background: The ability to create nanostructures with biomolecules is one of the key elements in nanobiotechnology. One of the problems is the expensive and mostly custom made equipment which is needed for their development. We intended to reduce material costs and aimed at miniaturization of the necessary tools that are essential for nanofabrication. Thus we combined the capabilities of molecular ink lithography with DNA-self-assembling capabilities to arrange DNA in an independent array which allows addressing molecules in nanoscale dimensions.
Results: For the construction of DNA based nanostructures a method is presented that allows an arrangement of DNA strands in such a way that they can form a grid that only depends on the spotted pattern of the anchor molecules. An atomic force microscope (AFM) has been used for molecular ink lithography to generate small spots. The sequential spotting process allows the immobilization of several different functional biomolecules with a single AFM-tip. This grid which delivers specific addresses for the prepared DNA-strand serves as a two-dimensional anchor to arrange the sequence according to the pattern. Once the DNA-nanoarray has been formed, it can be functionalized by PNA (peptide nucleic acid) to incorporate advanced structures.
Conclusions: The production of DNA-nanoarrays is a promising task for nanobiotechnology. The described method allows convenient and low cost preparation of nanoarrays. PNA can be used for complex functionalization purposes as well as a structural element.
The use of nano zerovalent iron (nZVI) for environmental remediation is a promising new technique for in situ remediation. Due to its high surface area and high reactivity, nZVI is able to dechlorinate organic contaminants and render them harmless. Limited mobility, due to fast aggregation and sedimentation of nZVI, limits the capability for source and plume remediation. Carbo-Iron is a newly developed material consisting of activated carbon particles (d50 = 0,8 µm) that are plated with nZVI particles. These particles combine the mobility of activated carbon and the reactivity of nZVI. This paper presents the first results of the transport experiments.
Public debate about energy relations between the EU and Russia is distorted. These distortions present considerable obstacles to the development of true partnership. At the core of the conflict is a struggle for resource rents between energy producing, energy consuming and transit countries. Supposed secondary aspects, however, are also of great importance. They comprise of geopolitics, market access, economic development and state sovereignty. The European Union, having engaged in energy market liberalisation, faces a widening gap between declining domestic resources and continuously growing energy demand. Diverse interests inside the EU prevent the definition of a coherent and respected energy policy. Russia, for its part, is no longer willing to subsidise its neighbouring economies by cheap energy exports. The Russian government engages in assertive policies pursuing Russian interests. In so far, it opts for a different globalisation approach, refusing the role of mere energy exporter. In view of the intensifying struggle for global resources, Russia, with its large energy potential, appears to be a very favourable option for European energy supplies, if not the best one. However, several outcomes of the strategic game between the two partners can be imagined. Engaging in non-cooperative strategies will in the end leave all stakeholders worse-off. The European Union should therefore concentrate on securing its partnership with Russia instead of damaging it. Stable cooperation would need the acceptance that the partner may pursue his own goals, which might be different from one’s own interests. The question is, how can a sustainable compromise be found? This thesis finds that a mix of continued dialogue, a tit for tat approach bolstered by an international institutional framework and increased integration efforts appears as a preferable solution.
Genetic variation is crucial for the long-term survival of the species as it provides the potential for adaptive responses to environmental changes such as emerging diseases. The Major Histocompatibility Complex (MHC) is a gene family that plays a central role in the vertebrate’s immune system by triggering the adaptive immune response after exposure to pathogens. MHC genes have become highly suitable molecular markers of adaptive significance. They synthesize two primary cell surface molecules namely MHC class I and class II that recognize short fragments of proteins derived respectively from intracellular (e.g. viruses) and extracellular (e.g. bacteria, protozoa, arthropods) origins and present them to immune cells. High levels of MHC polymorphism frequently observed in natural populations are interpreted as an adaptation to detect and present a wide array of rapidly evolving pathogens. This variation appears to be largely maintained by positive selection driven mainly by pathogenic selective pressures. For my doctoral research I focused on MHC I and II variation in free-ranging cheetahs (Acinonyx jubatus) and leopards (Panthera pardus) on Namibian farmlands. Both felid species are sympatric thus subject to similar pathogenic pressures but differ in their evolutionary and demographic histories. The main aims were to investigate 1) the extent and patterns of MHC variation at the population level in both felids, 2) the association between levels of MHC variation and disease resistance in free-ranging cheetahs, and 3) the role of selection at different time scales in shaping MHC variation in both felids. Cheetahs and leopards represent the largest free-ranging carnivores in Namibia. They concentrate in unprotected areas on privately owned farmlands where domestic and other wild animals also occur and the risk of pathogen transmission is increased. Thus, knowledge on adaptive genetic variation involved in disease resistance may be pertinent to both felid species’ conservation. The cheetah has been used as a classic example in conservation genetics textbooks due to overall low levels of genetic variation. Reduced variation at MHC genes has been associated with high susceptibility to infectious diseases in cheetahs. However, increased disease susceptibility has only been observed in captive cheetahs whereas recent studies in free-ranging Namibian cheetahs revealed a good health status. This raised the question whether the diversity at MHC I and II genes in free-ranging cheetahs is higher than previously reported. In this study, a total of 10 MHC I alleles and four MHC II alleles were observed in 149 individuals throughout Namibia. All alleles but one likely belong to functional MHC genes as their expression was confirmed. The observed alleles belong to four MHC I and three MHC II genes in the species as revealed by phylogenetic analyses. Signatures of historical positive selection acting on specific sites that interact directly with pathogen-derived proteins were detected in both MHC classes. Furthermore, a high genetic differentiation at MHC I was observed between Namibian cheetahs from east-central and north-central regions known to differ substantially in exposure to feline-specific viral pathogens. This suggests that the patterns of MHC I variation in the current population mirrors different pathogenic selective pressure imposed by viruses. Cheetahs showed low levels of MHC diversity compared with other mammalian species including felids, but this does not seem to influence the current immunocompetence of free-ranging cheetahs in Namibia and contradicts the previous conclusion that the cheetah is a paradigm species of disease susceptibility. However, it cannot be ruled out that the low MHC variation might limit a prosperous immunocompetence in the case of an emerging disease scenario because none of the remaining alleles might be able to recognize a novel pathogen. In contrast to cheetahs, leopards occur in most parts of Africa being perhaps the most abundant big cat in the continent. Leopards seem to have escaped from large-scale declines due to epizootics in the past in contrast to some free-ranging large carnivore populations in Africa that have been afflicted by epizootics. Currently, no information about the MHC sequence variation and constitution in African leopards exists. In this study, I characterized genetic variation at MHC I and MHC II genes in free-ranging leopards from Namibia. A total of six MHC I and six MHC II sequences were detected in 25 individuals from the east-central region. The maximum number of sequences observed per individual suggests that they likely correspond to at least three MHC I and three MHC II genes. Hallmarks of MHC evolution were confirmed such as historical positive selection, recombination and trans-species polymorphism. The low MHC variation detected in Namibian leopards is not conclusive and further research is required to assess the extent of MHC variation in different areas of its geographic range. Results from this thesis will contribute to better understanding the evolutionary significance of MHC and conservation implications in free-ranging felids. Translocation of wildlife is an increasingly used management tool for conservation purposes that should be conducted carefully as it may affect the ability of the translocated animals to cope with different pathogenic selective pressures.
AM symbiosis has a positive influence on plant P-nutrition and growth, but little is known about the molecular mechanism of the symbiosis adaptation to different phosphate conditions. The recently described induction of several pri-miR399 transcripts in mycorrhizal shoots and subsequent accumulation of mature miR399 in mycorrhizal roots indicates that local PHO2 expression must be controlled during symbiosis, presumably in order to sustain AM symbiosis development, in spite of locally increased Pi-concentration. A reverse genetic approach used in this study demonstrated that PHO2 and thus the PHR1-miR399-PHO2 signaling pathway, is involved in certain stages of progressive root colonization. In addition, a transcriptomic approach using a split-root system provided a comprehensive insight into the systemic transcriptional changes in mycorrhizal roots and shoots of M. truncatula in response to high phosphate conditions. With regard to the transcriptional responses of the root system, the results indicate that, although the colonization is drastically reduced, AM symbiosis is still functional at high Pi concentrations and might still be beneficial to the plant. Additionally, the data suggest that a specific root-borne mycorrhizal signal systemically induces protein synthesis, amino acid metabolism and photosynthesis at low Pi conditions, which is abolished at high Pi conditions. MiRNAs, such as miR399, are involved in long-distance signaling and are therefore potential systemic signals involved in AM symbiosis. A deep-sequencing approach identified 243 novel miRNAs in the root tissue of M. truncatula. Read-count analysis, qRT-PCR measurements and in situ hybridizations clearly indicated a regulation of miR5229a/b, miR5204, miR160f*, miR160c, miR169 and miR169d*/l*/m*/e.2* during arbuscular mycorrhizal symbiosis. Moreover, miR5204* represses a GRAS TF, which is specifically transcribed in mycorrhizal roots. Since miR5204* is induced by high Pi it might represent a further Pi status-mediating signal beside miR399. This study provides additional evidence that MtNsp2, a key regulator of symbiosis-signaling, is regulated and presumably spatially restricted by miR171h cleavage. In summary, a repression of mycorrhizal root colonization at high phosphate status is most likely due to a repression of the phosphate starvation responses and the loss of beneficial responses in mycorrhizal shoots. These findings provide a new basis for investigating the regulatory network leading to cellular reprogramming during interaction between plants, arbuscular mycorrhizal fungi and different phosphate conditions.
The northward movement and collision of the Arabian plate with Eurasia generates compressive stresses and resulting shortening in Iran. Within the Alborz Mountains, North Iran, a complex and not well understood system of strike-slip and thrust faults accomodates a fundamental part of the NNE-SSW oriented shortening. On 28th of May 2004 the Mw 6.3 Baladeh earthquake hit the north-central Alborz Mountains. It is one of the rare and large events in this region in modern time and thus a seldom chance to study earthquake mechanisms and the local ongoing deformation processes. It also demonstrated the high vulnerability of this densily populated region.
Building biological models by inferring functional dependencies from experimental data is an important issue in Molecular Biology. To relieve the biologist from this traditionally manual process, various approaches have been proposed to increase the degree of automation. However, available approaches often yield a single model only, rely on specific assumptions, and/or use dedicated, heuristic algorithms that are intolerant to changing circumstances or requirements in the view of the rapid progress made in Biotechnology. Our aim is to provide a declarative solution to the problem by appeal to Answer Set Programming (ASP) overcoming these difficulties. We build upon an existing approach to Automatic Network Reconstruction proposed by part of the authors. This approach has firm mathematical foundations and is well suited for ASP due to its combinatorial flavor providing a characterization of all models explaining a set of experiments. The usage of ASP has several benefits over the existing heuristic algorithms. First, it is declarative and thus transparent for biological experts. Second, it is elaboration tolerant and thus allows for an easy exploration and incorporation of biological constraints. Third, it allows for exploring the entire space of possible models. Finally, our approach offers an excellent performance, matching existing, special-purpose systems.
Indirect resource competition and interference are widely occurring mechanisms of interspecific interactions. We have studied the seasonal expression of these two interaction types within a two-species, boreal small mammal system. Seasons differ by resource availability, individual breeding state and intraspecific social system. Live-trapping methods were used to monitor space use and reproduction in 14 experimental populations of bank voles Myodes glareolus in large outdoor enclosures with and without a dominant competitor, the field vole Microtus agrestis. We further compared vole behaviour using staged dyadic encounters in neutral arenas in both seasons. Survival of the non-breeding overwintering bank voles was not affected by competition. In the spring, the numbers of male bank voles, but not of females, were reduced significantly in the competition populations. Bank vole home ranges expanded with vole density in the presence of competitors, indicating food limitation. A comparison of behaviour between seasons based on an analysis of similarity revealed an avoidance of costly aggression against opponents, independent of species. Interactions were more aggressive during the summer than during the winter, and heterospecific encounters were more aggressive than conspecific encounters. Based on these results, we suggest that interaction types and their respective mechanisms are not either–or categories and may change over the seasons. During the winter, energy constraints and thermoregulatory needs decrease direct aggression, but food constraints increase indirect resource competition. Direct interference appears in the summer, probably triggered by each individual’s reproductive and hormonal state and the defence of offspring against conspecific and heterospecific intruders. Both interaction forms overlap in the spring, possibly contributing to spring declines in the numbers of subordinate species.
This essay revisits Ian McEwan’s extremely successful novel Saturday, and interrogates its exemplary assessment of the British cultural climate after 9/11. The particular focus is on McEwan’s extensive recourse to the writings of Matthew Arnold, whose melancholy outlook on culture and anarchy McEwan basically translates into the 21st century without much ideological fraction. This relapse into Victorian liberal humanism as consolation for a Western world besieged by the contingencies of terrorism is extremely problematic. Not only does it wilfully ignore the transcultural realities of modern Britain, it also promotes an ahistorical and apolitical mode of critical inquiry which may be called reductive at best in view of the global challenges that the novel addresses.
Fusarium spp. infection of cereal grain is a common problem, which leads to a dramatic loss of grain quality. The aim of the present study was to investigate the effect of Fusarium infection on the wheat storage protein gluten and its fractions, the gliadins and glutenins, in an in vitro model system. Gluten proteins were digested by F. graminearum proteases for 2, 4, 8 and 24 h, separated by Osborne fractionation and characterised by chromatographic (RP-HPLC) and electrophoretic analysis (SDS-Page). Gluten digestion by F. graminearum proteases showed in comparison with gliadins a preference for the glutenins whereas the HMW subfraction was at most affected. In comparison with a untreated control, the HMW subfraction was degraded of about 97% after 4 h incubation with Fusarium proteases. Separate digestion of gliadin and glutenin underlined the preference for HMW-GS. Analogue to the observed change in the gluten composition, the yield of the proteins extracted changed. A higher amount of glutenin fragments was found in the gliadin extraction solution after digestion and could mask a gliadin destruction at the same time. This observation can contribute to explain the frequently reported reduced glutenin amount parallel to an increase in gliadin quantity after Fusarium infection in grains.
Urbanity and literature
(2011)
Transarea studies focus upon spaces as created by the movements that criss-cross them. From this point of view, from its very beginnings, literature is closely interrelated with a vectorial (and much less with a purely spatial) conception of history - and with urbanity, which plays a decisive role in Gilgamesh's travels through a (narrative) cosmos centered upon the city of Uruk. This article explores the city as a transareal space of movement in three examples of literature, with no fixed abode, around the turn of the millennium, i.e. Assia Djebar's Les Nuits de Strasbourg, Emine Sevgi Oezdamar's Istanbul-Berlin Trilogy, and Cecile Wajsbrot's L'ile aux musees. These three writers project, in a very specific way, cities in motion as anagrammatic and fractal structures.
Ground-based gamma-ray astronomy has had a major breakthrough with the impressive results obtained using systems of imaging atmospheric Cherenkov telescopes. Ground-based gamma-ray astronomy has a huge potential in astrophysics, particle physics and cosmology. CTA is an international initiative to build the next generation instrument, with a factor of 5-10 improvement in sensitivity in the 100 GeV-10 TeV range and the extension to energies well below 100 GeV and above 100 TeV. CTA will consist of two arrays (one in the north, one in the south) for full sky coverage and will be operated as open observatory. The design of CTA is based on currently available technology. This document reports on the status and presents the major design concepts of CTA.
In this work, the development of a new molecular building block, based on synthetic peptides derived from decorin, is presented. These peptides represent a promising basis for the design of polymer-based biomaterials that mimic the ECM on a molecular level and exploit specific biological recognition for technical applications. Multiple sequence alignments of the internal repeats of decorin that formed the inner and outer surface of the arch-shaped protein were used to develop consensus sequences. These sequences contained conserved sequence motifs that are likely to be related to structural and functional features of the protein. Peptides representative for the consensus sequences were synthesized by microwave-assisted solid phase peptide synthesis and purified by RP-HPLC, with purities higher than 95 mol%. After confirming the desired masses by MALDI-TOF-MS, the primary structure of each peptide was investigated by 1H and 2D NMR, from which a full assignment of the chemical shifts was obtained. The characterization of the peptides conformation in solution was performed by CD spectroscopy, which demonstrated that using TFE, the peptides from the outer surface of decorin show a high propensity to fold into helical structures as observed in the original protein. To the contrary, the peptides from the inner surface did not show propensity to form stable secondary structure. The investigation of the binding capability of the peptides to Collagen I was performed by surface plasmon resonance analyses, from which all but one of the peptides representing the inner surface of decorin showed binding affinity to collagen with values of dissociation constant between 2•10-7 M and 2.3•10-4 M. On the other hand, the peptides representative for the outer surface of decorin did not show any significant interaction to collagen. This information was then used to develop experimental demonstration for the binding capabilities of the peptides from the inner surface of decorin to collagen even when used in more complicated situations close to possible appications. With this purpose, the peptide (LRELHLNNN) which showed the highest binding affinity to collagen (2•10-7 M) was functionalized with an N-terminal triple bond in order to obtain a peptide dimer via copper(I)-catalyzed cycloaddition reaction with 4,4'-diazidostilbene-2,2'-disulfonic acid. Rheological measurements showed that the presence of the peptide dimer was able to enhance the elastic modulus (G') of a collagen gel from ~ 600 Pa (collagen alone) to ~ 2700 Pa (collagen and peptide dimer). Moreover, it was shown that the mechanical properties of a collagen gel can be tailored by using different molar ratios of peptide dimer respect to collagen. The same peptide, functionalized with the triple bond, was used to obtain a peptide-dye conjugate by coupling it with N-(5'-azidopentanoyl)-5-aminofluorescein. An aqueous solution (5 vol% methanol) of the peptide dye conjugate was injected into a collagen and a hyaluronic acid (HA) gel and images of fluorescence detection showed that the diffusion of the peptide was slower in the collagen gel compared to the HA gel. The third experimental demonstration was gained using the peptide (LSELRLHNN) which showed the lower binding affinity (2.3•10-4 M) to collagen. This peptide was grafted to hyaluronic acid via EDC-chemistry, with a degree of functionalization of 7 ± 2 mol% as calculated by 1H-NMR. The grafting was further confirmed by FTIR and TGA measurements, which showed that the onset of decomposition for the HA-g-peptide decreased by 10 °C compared to the native HA. Rheological measurements showed that the elastic modulus of a system based on collagen and HA-g-peptide increased by almost two order of magnitude (G' = 200 Pa) compared to a system based on collagen and HA (G' = 0.9 Pa). Overall, this study showed that the synthetic peptides, which were identified from decorin, can be applied as potential building blocks for biomimetic materials that function via biological recognition.
rezensiertes Werk: Marx, Peter W.:Ein theatralisches Zeitalter : Bürgerliche Selbstinszenierungen um 1900. - Tübingen [u.a.] : A. Francke, 2008. - 429 S. ISBN 978-3-772-08220-7
The needs for sustainable energy generation, but also a sustainable chemistry display the basic motivation of the current thesis. By different single investigated cases, which are all related to the element carbon, the work can be devided into two major topics. At first, the sustainable synthesis of “useful” carbon materials employing the process of hydrothermal carbonisation (HC) is described. In the second part, the synthesis of heteroatom - containing carbon materials for electrochemical and fuel cell applications employing ionic liquid precursors is presented. On base of a thorough review of the literature on hydrothermolysis and hydrothermal carbonisation of sugars in addition to the chemistry of hydroxymethylfurfural, mechanistic considerations of the formation of hydrothermal carbon are proposed. On the base of these reaction schemes, the mineral borax, is introduced as an additive for the hydrothermal carbonisation of glucose. It was found to be a highly active catalyst, resulting in decreased reaction times and increased carbon yields. The chemical impact of borax, in the following is exploited for the modification of the micro- and nanostructure of hydrothermal carbon. From the borax - mediated aggregation of those primary species, widely applicable, low density, pure hydrothermal carbon aerogels with high porosities and specific surface areas are produced. To conclude the first section of the thesis, a short series of experiments is carried out, for the purpose of demonstrating the applicability of the HC model to “real” biowaste i.e. watermelon waste as feedstock for the production of useful materials. In part two cyano - containing ionic liquids are employed as precursors for the synthesis of high - performance, heteroatom - containing carbon materials. By varying the ionic liquid precursor and the carbonisation conditions, it was possible to design highly active non - metal electrocatalyst for the reduction of oxygen. In the direct reduction of oxygen to water (like used in polymer electrolyte fuel cells), compared to commercial platinum catalysts, astonishing activities are observed. In another example the selective and very cost efficient electrochemical synthesis of hydrogen peroxide is presented. In a last example the synthesis of graphitic boron carbon nitrides from the ionic liquid 1 - Ethyl - 3 - methylimidazolium - tetracyanoborate is investigated in detail. Due to the employment of unreactive salts as a new tool to generate high surface area these materials were first time shown to be another class of non - precious metal oxygen reduction electrocatalyst.
This is the second part of the presentation of data elicited by means of QUIS within the project on information structure in Gur and Kwa languages. Whereas the first part (Anne Schwarz) introduces the project and the rationals behind the development of the focus translation task, this part provides some comparative remarks gained from the data presented in both parts.
A key non-destructive technique for analysis, optimization and developing of new functional materials such as sensors, transducers, electro-optical and memory devices is presented. The Thermal-Pulse Tomography (TPT) provides high-resolution three-dimensional images of electric field and polarization distribution in a material. This thermal technique use a pulsed heating by means of focused laser light which is absorbed by opaque electrodes. The diffusion of the heat causes changes in the sample geometry, generating a short-circuit current or change in surface potential, which contains information about the spatial distribution of electric dipoles or space charges. Afterwards, a reconstruction of the internal electric field and polarization distribution in the material is possible via Scale Transformation or Regularization methods. In this way, the TPT was used for the first time to image the inhomogeneous ferroelectric switching in polymer ferroelectric films (candidates to memory devices). The results shows the typical pinning of electric dipoles in the ferroelectric polymer under study and support the previous hypotheses of a ferroelectric reversal at a grain level via nucleation and growth. In order to obtain more information about the impact of the lateral and depth resolution of the thermal techniques, the TPT and its counterpart called Focused Laser Intensity Modulation Method (FLIMM) were implemented in ferroelectric films with grid-shaped electrodes. The results from both techniques, after the data analysis with different regularization and scale methods, are in total agreement. It was also revealed a possible overestimated lateral resolution of the FLIMM and highlights the TPT method as the most efficient and reliable thermal technique. After an improvement in the optics, the Thermal-Pulse Tomography method was implemented in polymer-dispersed liquid crystals (PDLCs) films, which are used in electro-optical applications. The results indicated a possible electrostatic interaction between the COH group in the liquid crystals and the fluorinate atoms of the used ferroelectric matrix. The geometrical parameters of the LC droplets were partially reproduced as they were compared with Scanning Electron Microscopy (SEM) images. For further applications, it is suggested the use of a non-strong-ferroelectric polymer matrix. In an effort to develop new polymerferroelectrets and for optimizing their properties, new multilayer systems were inspected. The results of the TPT method showed the non-uniformity of the internal electric-field distribution in the shaped-macrodipoles and thus suggested the instability of the sample. Further investigation on multilayers ferroelectrets was suggested and the implementation of less conductive polymers layers too.
This article presents several acceptability rating experiments concerned with crossing wh-movement in German multiple questions. Our results show that there is no general superiority effect in German, thus refuting claims to the contrary by Featherston (2005). However, acceptability is reduced when a whphrase crosses a wh-subject with which it agrees in animacy. We explain this finding in terms of the availability of different sorting keys for the answers to the multiple questions.
Answer Set Programming (ASP) is an emerging paradigm for declarative programming, in which a computational problem is specified by a logic program such that particular models, called answer sets, match solutions. ASP faces a growing range of applications, demanding for high-performance tools able to solve complex problems. ASP integrates ideas from a variety of neighboring fields. In particular, automated techniques to search for answer sets are inspired by Boolean Satisfiability (SAT) solving approaches. While the latter have firm proof-theoretic foundations, ASP lacks formal frameworks for characterizing and comparing solving methods. Furthermore, sophisticated search patterns of modern SAT solvers, successfully applied in areas like, e.g., model checking and verification, are not yet established in ASP solving. We address these deficiencies by, for one, providing proof-theoretic frameworks that allow for characterizing, comparing, and analyzing approaches to answer set computation. For another, we devise modern ASP solving algorithms that integrate and extend state-of-the-art techniques for Boolean constraint solving. We thus contribute to the understanding of existing ASP solving approaches and their interconnections as well as to their enhancement by incorporating sophisticated search patterns. The central idea of our approach is to identify atomic as well as composite constituents of a propositional logic program with Boolean variables. This enables us to describe fundamental inference steps, and to selectively combine them in proof-theoretic characterizations of various ASP solving methods. In particular, we show that different concepts of case analyses applied by existing ASP solvers implicate mutual exponential separations regarding their best-case complexities. We also develop a generic proof-theoretic framework amenable to language extensions, and we point out that exponential separations can likewise be obtained due to case analyses on them. We further exploit fundamental inference steps to derive Boolean constraints characterizing answer sets. They enable the conception of ASP solving algorithms including search patterns of modern SAT solvers, while also allowing for direct technology transfers between the areas of ASP and SAT solving. Beyond the search for one answer set of a logic program, we address the enumeration of answer sets and their projections to a subvocabulary, respectively. The algorithms we develop enable repetition-free enumeration in polynomial space without being intrusive, i.e., they do not necessitate any modifications of computations before an answer set is found. Our approach to ASP solving is implemented in clasp, a state-of-the-art Boolean constraint solver that has successfully participated in recent solver competitions. Although we do here not address the implementation techniques of clasp or all of its features, we present the principles of its success in the context of ASP solving.
Preference handling and optimization are indispensable means for addressing nontrivial applications in Answer Set Programming (ASP). However, their implementation becomes difficult whenever they bring about a significant increase in computational complexity. As a consequence, existing ASP systems do not offer complex optimization capacities, supporting, for instance, inclusion-based minimization or Pareto efficiency. Rather, such complex criteria are typically addressed by resorting to dedicated modeling techniques, like saturation. Unlike the ease of common ASP modeling, however, these techniques are rather involved and hardly usable by ASP laymen. We address this problem by developing a general implementation technique by means of meta-prpogramming, thus reusing existing ASP systems to capture various forms of qualitative preferences among answer sets. In this way, complex preferences and optimization capacities become readily available for ASP applications.
Using the notion of an elementary loop, Gebser and Schaub (2005. Proceedings of the Eighth International Conference on Logic Programming and Nonmonotonic Reasoning (LPNMR’05 ), 53–65) refined the theorem on loop formulas attributable to Lin and Zhao (2004) by considering loop formulas of elementary loops only. In this paper, we reformulate the definition of an elementary loop, extend it to disjunctive programs, and study several properties of elementary loops, including how maximal elementary loops are related to minimal unfounded sets. The results provide useful insights into the stable model semantics in terms of elementary loops. For a nondisjunctive program, using a graph-theoretic characterization of an elementary loop, we show that the problem of recognizing an elementary loop is tractable. On the other hand, we also show that the corresponding problem is coNP-complete for a disjunctive program. Based on the notion of an elementary loop, we present the class of Head-Elementary-loop-Free (HEF) programs, which strictly generalizes the class of Head-Cycle-Free (HCF) programs attributable to Ben-Eliyahu and Dechter (1994. Annals of Mathematics and Artificial Intelligence 12, 53–87). Like an HCF program, an HEF program can be turned into an equivalent nondisjunctive program in polynomial time by shifting head atoms into the body.
We introduce an approach to detecting inconsistencies in large biological networks by using answer set programming. To this end, we build upon a recently proposed notion of consistency between biochemical/genetic reactions and high-throughput profiles of cell activity. We then present an approach based on answer set programming to check the consistency of large-scale data sets. Moreover, we extend this methodology to provide explanations for inconsistencies by determining minimal representations of conflicts. In practice, this can be used to identify unreliable data or to indicate missing reactions.
The armed conflict in Afghanistan since 2001 has raised manifold questions pertaining to the humanitarian rules relative to the conduct of hostilities. In Afghanistan, as is often the case in so-called asymmetric conflicts, the geographical and temporal boundaries of the battlefield, and the distinction between civilians and fighters, are increasingly blurred. As a result, the risks for both civilians and soldiers operating in Afghanistan are high. The objective of this article is to assess whether - and if so how much - the armed conflict in Afghanistan has affected the application and interpretation of the principles of distinction, proportionality, and precaution - principles that form the core of legal rules pertaining to the conduct of hostilities.
During reading oculomotor processes guide the eyes over the text. The visual information recorded is accessed, evaluated and processed. Only by retrieving the meaning of a word from the long-term memory, as well as through the connection and storage of the information about each individual word, is it possible to access the semantic meaning of a sentence. Therefore memory, and here in particular working memory, plays a pivotal role in the basic processes of reading. The following dissertation investigates to what extent different demands on memory and memory capacity have an effect on eye movement behavior while reading. The frequently used paradigm of the reading span task, in which test subjects read and evaluate individual sentences, was used for the experimental review of the research questions. The results speak for the fact that working memory processes have a direct effect on various eye movement measurements. Thus a high working memory load, for example, reduced the perceptual span while reading. The lower the individual working memory capacity of the reader was, the stronger was the influence of the working memory load on the processing of the sentence.
This article presents a situation description production experiment investigating the interaction between syntax and information structure in Akan, a tone language that belongs to the Kwa branch of the Niger- Congo family spoken in Ghana. Information structure was elicited via context questions that put the object in narrow informational focus or narrow corrective focus while controlling for the tonal structure of the target word. Contrary to the prediction that corrective focus is marked by fronting and morphological marking of the focused constituent the data suggest that the in-situ strategy is the preferred one.
Resolute readings of later Wittgenstein and the challenge of avoiding hierarchies in philosophy
(2011)
This dissertation addresses the question: How did later Wittgenstein aim to achieve his goal of putting forward a way of dissolving philosophical problems which centered on asking ourselves what we mean by our words – yet which did not entail any claims about the essence of language and meaning? This question is discussed with reference to “resolute” readings of Wittgenstein. I discuss the readings of James Conant, Oskari Kuusela, and Martin Gustafsson. I follow Oskari Kuusela’s claim that in order to fully appreciate how later Wittgenstein meant to achieve his goal, we need to clearly see how he aimed to do away with hierarchies in philosophy: Not only is the dissolution of philosophical problems via the method of clarifying the grammar of expressions to be taken as independent from any theses about what meaning must be – but furthermore, it is to be taken as independent from the dissolution of any particular problem via this method. As Kuusela stresses, this also holds for the problems involving rule-following and meaning: the clarification of the grammar of “rule” and “meaning” has no foundational status – it is nothing on which the method of clarifying the grammar of expressions as such were meant to in any way rely on. The lead question of this dissertation then is: What does it mean to come to see that the method of dissolving philosophical problems by asking “How is this word actually used?” does not in any way rely on the results of our having investigated the grammar of the particular concepts “rule” and “meaning”? What is the relation of such results – results such as “To follow a rule, [...], to obey an order, [...] are customs (uses, institutions)” or “The meaning of a word is its use in the language” – to this method? From this vantage point, I concern myself with two aspects of the readings of Gustafsson and Kuusela. In Gustafsson, I concern myself with his idea that the dissolution of philosophical problems in general “relies on” the very agreement which – during the dissolution of the rule-following problem – comes out as a presupposition for our talk of “meaning” in terms of rules. In Kuusela, I concern myself with his idea that Wittgenstein, in adopting a way of philosophical clarification which investigates the actual use of expressions, is following the model of “meaning as use” – which model he had previously introduced in order to perspicuously present an aspect of the actual use of the word “meaning”. This dissertation aims to show how these two aspects of Gustafsson’s and Kuusela’s readings still fail to live up to the vision of Wittgenstein as a philosopher who aimed to do away with any hierarchies in philosophy. I base this conclusion on a detailed analysis of which of the occasions where Wittgenstein invokes the notions of “use” and “application” (as also “agreement”) have to do with the dissolution of a specific problem only, and which have to do with the dissolution of philosophical problems in general. I discuss Wittgenstein’s remarks on rule-following, showing how in the dissolution of the rule-following paradox, notions such as “use”, “application”, and “practice” figure on two distinct logical levels. I then discuss an example of what happens when this distinction is not duly heeded: Gordon Baker and Peter Hacker’s idea that the rule-following remarks have a special significance for his project of dissolving philosophical problems as such. I furnish an argument to the effect that their idea that the clarification of the rules of grammar of the particular expression “following a rule” could answer a question about rules of grammar in general rests on a conflation of the two logical levels on which “use” occurs in the rule-following remarks, and that it leads into a regress. I then show that Gustafsson’s view – despite its decisive advance over Baker and Hacker – contains a version of that same idea, and that it likewise leads into a regress. Finally, I show that Kuusela’s idea of a special significance of the model “meaning as use” for the whole of the method of stating rules for the use of words is open to a regress argument of a similar kind as that he himself advances against Baker and Hacker. I conclude that in order to avoid such a regress, we need to reject the idea that the grammatical remark “The meaning of a word is its use in the language” – because of the occurrence of “use” in it – stood in any special relation to the method of dissolving philosophical problems by describing the use of words. Rather, we need to take this method as independent from this outcome of the investigation of the use of the particular word “meaning”.
Regulation of gene transcription plays a major role in mediating cellular responses and physiological behavior in all known organisms. The finding that similar genes are often regulated in a similar manner (co-regulated or "co-expressed") has directed several "guilt-by-association" approaches in order to reverse-engineer the cellular transcriptional networks using gene expression data as a compass. This kind of studies has been considerably assisted in the recent years by the development of high-throughput transcript measurement platforms, specifically gene microarrays and next-generation sequencing. In this thesis, I describe several approaches for improving the extraction and interpretation of the information contained in microarray based gene expression data, through four steps: (1) microarray platform design, (2) microarray data normalization, (3) gene network reverse engineering based on expression data and (4) experimental validation of expression-based guilt-by-association inferences. In the first part test case is shown aimed at the generation of a microarray for Thellungiella salsuginea, a salt and drought resistant close relative to the model plant Arabidopsis thaliana; the transcripts of this organism are generated on the combination of publicly available ESTs and newly generated ad-hoc next-generation sequencing data. Since the design of a microarray platform requires the availability of highly reliable and non-redundant transcript models, these issues are addressed consecutively, proposing several different technical solutions. In the second part I describe how inter-array correlation artifacts are generated by the common microarray normalization methods RMA and GCRMA, together with the technical and mathematical characteristics underlying the problem. A solution is proposed in the form of a novel normalization method, called tRMA. The third part of the thesis deals with the field of expression-based gene network reverse engineering. It is shown how different centrality measures in reverse engineered gene networks can be used to distinguish specific classes of genes, in particular essential genes in Arabidopsis thaliana, and how the use of conditional correlation can add a layer of understanding over the information flow processes underlying transcript regulation. Furthermore, several network reverse engineering approaches are compared, with a particular focus on the LASSO, a linear regression derivative rarely applied before in global gene network reconstruction, despite its theoretical advantages in robustness and interpretability over more standard methods. The performance of LASSO is assessed through several in silico analyses dealing with the reliability of the inferred gene networks. In the final part, LASSO and other reverse engineering methods are used to experimentally identify novel genes involved in two independent scenarios: the seed coat mucilage pathway in Arabidopsis thaliana and the hypoxic tuber development in Solanum tuberosum. In both cases an interesting method complementarity is shown, which strongly suggests a general use of hybrid approaches for transcript expression-based inferences. In conclusion, this work has helped to improve our understanding of gene transcription regulation through a better interpretation of high-throughput expression data. Part of the network reverse engineering methods described in this thesis have been included in a tool (CorTo) for gene network reverse engineering and annotated visualization from custom transcription datasets.
Ngizim fieldnotes
(2011)
This chapter presents field notes of the West Chadic language Ngizim, spoken in North-East Nigeria. In Ngizim, subject focus is indicated by subject inversion, whereas the word order of sentences with focused non-subjects can remain unchanged. The goal of the field work was to find out more about focus marking in Ngizim.
Soil moisture is a key state variable that controls runoff formation, infiltration and partitioning of radiation into latent and sensible heat. However, the experimental characterisation of near surface soil moisture patterns and their controls on runoff formation remains a challenge. This subject was one aspect of the BMBF-funded OPAQUE project (operational discharge and flooding predictions in head catchments). As part of that project the focus of this dissertation is on: (1) testing the methodology and feasibility of the Spatial TDR technology in producing soil moisture profiles along TDR probes, including an inversion technique of the recorded signal in heterogeneous field soils, (2) the analysis of spatial variability and temporal dynamics of soil moisture at the field scale including field experiments and hydrological modelling, (3) the application of models of different complexity for understanding soil moisture dynamics and its importance for runoff generation as well as for improving the prediction of runoff volumes. To fulfil objective 1, several laboratory experiments were conducted to understand the influence of probe rod geometry and heterogeneities in the sampling volume under different wetness conditions. This includes a detailed analysis on how these error sources affect retrieval of soil moisture profiles in soils. Concerning objective 2 a sampling strategy of two TDR clusters installed in the head water of the Wilde Weißeritz catchment (Eastern Ore Mountains, Germany) was used to investigate how well “the catchment state” can be characterised by means of distributed soil moisture data observed at the field scale. A grassland site and a forested site both located on gentle slopes were instrumented with two Spatial TDR clusters that consist of up to 39 TDR probes. Process understanding was gained by modelling the interaction of evapotranspiration and soil moisture with the hydrological process model CATFLOW. A field scale irrigation experiment was carried out to investigate near subsurface processes at the hillslope scale. The interactions of soil moisture and runoff formation were analysed using discharge data from three nested catchments: the Becherbach with a size of 2 km², the Rehefeld catchment (17 km²) and the superordinate Ammelsdorf catchment (49 km²). Statistical analyses including observations of pre-event runoff, soil moisture and different rainfall characteristics were employed to predict stream flow volume. On the different scales a strong correlation between the average soil moisture and the runoff coefficients of rainfall-runoff events could be found, which almost explains equivalent variability as the pre-event runoff. Furthermore, there was a strong correlation between surface soil moisture and subsurface wetness with a hysteretic behaviour between runoff soil moisture. To fulfil objective 3 these findings were used in a generalised linear model (GLM) analysis which combines state variables describing the catchments antecedent wetness and variables describing the meteorological forcing in order to predict event runoff coefficients. GLM results were compared to simulations with the catchment model WaSiM ETH. Hereby were the model results of the GLMs always better than the simulations with WaSiM ETH. The GLM analysis indicated that the proposed sampling strategy of clustering TDR probes in typical functional units is a promising technique to explore soil moisture controls on runoff generation and can be an important link between the scales. Long term monitoring of such sites could yield valuable information for flood warning and forecasting by identifying critical soil moisture conditions for the former and providing a better representation of the initial moisture conditions for the latter.
We present XMM-Newton and Chandra observations of the born-again planetary nebula A 30. These X-ray observations reveal a bright unresolved source at the position of the central star whose X-ray luminosity exceeds by far the model expectations for photospheric emission and for shocks within the stellar wind. We suggest that a “born-again hot bubble” may be responsible for this X-ray emission. Diffuse X-ray emission associated with the petal-like features and one of the H-poor knots seen in the optical is also found. The weakened emission of carbon lines in the spectrum of the diffuse emission can be interpreted as the dilution of stellar wind by mass-loading or as the detection of material ejected during a very late thermal pulse.
Within this work, three physicochemical methods for the hydrophobization of initially hydrophilic solid particles are investigated. The modified particles are then used for the stabilization of oil-in-water (o/w) emulsions. For all introduced methods electrostatic interactions between strongly or weakly charged groups in the system are es-sential. (i) Short chain alkylammonium bromides (C4 – C12) adsorb on oppositely charged solid particles. Macroscopic contact angle measurements of water droplets under air and hexane on flat silica surfaces in dependency of the surface charge density and alkylchain-length allow the calculation of the surface energy and give insights into the emulsification properties of solid particles modified with alkyltrimethylammonium bromides. The measure-ments show an increase of the contact angle with increasing surface charge density, due to the enhanced adsorp-tion of the oppositely charged alkylammonium bromides. Contact angles are higher for longer alkylchain lengths. The surface energy calculations show that in particular the surface-hexane or surface-air interfacial en-ergy is being lowered upon alkylammonium adsorption, while a significant increase of the surface-water interfa-cial energy occurs only at long alkyl chain lengths and high surface charge densities. (ii) The thickness and the charge density of an adsorbed weak polyelectrolyte layer (e.g. PMAA, PAH) influence the wettability of nanoparticles (e.g. alumina, silica, see Scheme 1(b)). Furthermore, the isoelectric point and the pH range of colloidal stability of particle-polyelectrolyte composites depend on the thickness of the weak polye-lectrolyte layer. Silica nanoparticles with adsorbed PAH and alumina nanoparticles with adsorbed PMAA be-come interfacially active and thus able to stabilize o/w emulsions when the degree of dissociation of the polye-lectrolyte layer is below 80 %. The average droplet size after emulsification of dodecane in water depends on the thickness and the degree of dissociation of the adsorbed PE-layer. The visualization of the particle-stabilized o/w emulsions by cryogenic SEM shows that for colloidally stable alumina-PMAA composites the oil-water interface is covered with a closely packed monolayer of particles, while for the colloidally unstable case closely packed aggregated particles deposit on the interface. (iii) By emulsifying a mixture of the corrosion inhibitor 8-hydroxyquinoline (8-HQ) and styrene with silica nanoparticles a highly stable o/w emulsion can be obtained in a narrow pH window. The amphoteric character of 8-HQ enables a pH dependent electrostatic interaction with silica nanoparticles, which can render them interfa-cially active. Depending on the concentration and the degree of dissociation of 8-HQ the adsorption onto silica results from electrostatic or aromatic interactions between 8-HQ and the particle-surface. At intermediate amounts of adsorbed 8-HQ the oil wettability of the particles becomes sufficient for stabilizing o/w emulsions. Cryogenic SEM visualization shows that the particles arrange then in a closely packed shell consisting of partly of aggregated domains on the droplet interface. For further increasing amounts of adsorbed 8-HQ the oil wet-tability is reduced again and the particles ability to stabilize emulsions decreases. By the addition of hexadecane to the oil phase the size of the droplets can be reduced down to 200 nm by in-creasing the silica mass fraction. Subsequent polymerization produces corrosion inhibitor filled (20 wt-%) poly-styrene-silica composite particles. The measurement of the release of 8-hydroxyquinoline shows a rapid increase of 8-hydroxyquinoline in a stirred aqueous solution indicating the release of the total content in less than 5 min-utes. The method is extended for the encapsulation of other organic corrosion inhibitors. The silica-polymer-inhibitor composite particles are then dispersed in a water based alkyd emulsion, and the dispersion is used to coat flat aluminium substrates. After drying and cross-linking the polmer-film Confocal Laser Scanning Micros-copy is employed revealing a homogeneous distribution of the particles in the film. Electrochemical Impedance Spectroscopy in aqueous electrolyte solutions shows that films with aggregated particle domains degrade with time and don’t provide long-term corrosion protection of the substrate. However, films with highly dispersed particles have high barrier properties for corrosive species. The comparison of films containing silica-polystyrene composite particles with and without 8-hydroxyquinoline shows higher electrochemical impedances when the inhibitor is present in the film. By applying the Scanning Vibrating Electrode Technique the localized corrosion rate in the fractured area of scratched polymer films containing the silica-polymer-inhibitor composite particles is studied. Electrochemical corrosion cannot be suppressed but the rate is lowered when inhibitor filled composite particles are present in the film. By depositing six polyelectrolyte layers on particle stabilized emulsion droplets their surface morphology changes significantly as shown by SEM visualization. When the oil wettability of the outer polyelectrolyte layer increases, the polyelectrolyte coated droplets can act as emulsion stabilizers themselves by attaching onto bigger oil droplets in a closely packed arrangement. In the presence of 3 mM LaCl3 8-HQ hydrophobized silica particles aggregate strongly on the oil-water inter-face. The application of an ultrasonic field can remove two dimensional shell-compartments from the droplet surface, which are then found in the aqueous bulk phase. Their size ranges up to 1/4th of the spherical particle shell.
Earthquake faults interact with each other in many different ways and hence earthquakes cannot be treated as individual independent events. Although earthquake interactions generally lead to a complex evolution of the crustal stress field, it does not necessarily mean that the earthquake occurrence becomes random and completely unpredictable. In particular, the interplay between earthquakes can rather explain the occurrence of pronounced characteristics such as periods of accelerated and depressed seismicity (seismic quiescence) as well as spatiotemporal earthquake clustering (swarms and aftershock sequences). Ignoring the time-dependence of the process by looking at time-averaged values – as largely done in standard procedures of seismic hazard assessment – can thus lead to erroneous estimations not only of the activity level of future earthquakes but also of their spatial distribution. Therefore, it exists an urgent need for applicable time-dependent models. In my work, I aimed at better understanding and characterization of the earthquake interactions in order to improve seismic hazard estimations. For this purpose, I studied seismicity patterns on spatial scales ranging from hydraulic fracture experiments (meter to kilometer) to fault system size (hundreds of kilometers), while the temporal scale of interest varied from the immediate aftershock activity (minutes to months) to seismic cycles (tens to thousands of years). My studies revealed a number of new characteristics of fluid-induced and stress-triggered earthquake clustering as well as precursory phenomena in earthquake cycles. Data analysis of earthquake and deformation data were accompanied by statistical and physics-based model simulations which allow a better understanding of the role of structural heterogeneities, stress changes, afterslip and fluid flow. Finally, new strategies and methods have been developed and tested which help to improve seismic hazard estimations by taking the time-dependence of the earthquake process appropriately into account.
In this thesis, we discuss the formulation of variational problems on supermanifolds. Supermanifolds incorporate bosonic as well as fermionic degrees of freedom. Fermionic fields take values in the odd part of an appropriate Grassmann algebra and are thus showing an anticommutative behaviour. However, a systematic treatment of these Grassmann parameters requires a description of spaces as functors, e.g. from the category of Grassmann algberas into the category of sets (or topological spaces, manifolds). After an introduction to the general ideas of this approach, we use it to give a description of the resulting supermanifolds of fields/maps. We show that each map is uniquely characterized by a family of differential operators of appropriate order. Moreover, we demonstrate that each of this maps is uniquely characterized by its component fields, i.e. by the coefficients in a Taylor expansion w.r.t. the odd coordinates. In general, the component fields are only locally defined. We present a way how to circumvent this limitation. In fact, by enlarging the supermanifold in question, we show that it is possible to work with globally defined components. We eventually use this formalism to study variational problems. More precisely, we study a super version of the geodesic and a generalization of harmonic maps to supermanifolds. Equations of motion are derived from an energy functional and we show how to decompose them into components. Finally, in special cases, we can prove the existence of critical points by reducing the problem to equations from ordinary geometric analysis. After solving these component equations, it is possible to show that their solutions give rise to critical points in the functor spaces of fields.
CSOM/PL is a software product line (SPL) derived from applying multi-dimensional separation of concerns (MDSOC) techniques to the domain of high-level language virtual machine (VM) implementations. For CSOM/PL, we modularised CSOM, a Smalltalk VM implemented in C, using VMADL (virtual machine architecture description language). Several features of the original CSOM were encapsulated in VMADL modules and composed in various combinations. In an evaluation of our approach, we show that applying MDSOC and SPL principles to a domain as complex as that of VMs is not only feasible but beneficial, as it improves understandability, maintainability, and configurability of VM implementations without harming performance.
Although horses and donkeys belong to the same genus, their genetic characteristics probably result in specific proteomes and post-translational modifications (PTM) of proteins. Since PTM can alter protein properties, specific PTM may contribute to species-specific characteristics. Therefore, the aim of the present study was to analyse differences in serum protein profiles of horses and donkeys as well as mules, which combine the genetic backgrounds of both species. Additionally, changes in PTM of the protein transthyretin (TTR) were analysed. Serum protein profiles of each species (five animals per species) were determined using strong anion exchanger ProteinChips (R) (Bio-Rad, Munich, Germany) in combination with surface-enhanced laser desorption ionisation-time of flight MS. The PTM of TTR were analysed subsequently by immunoprecipitation in combination with matrix-assisted laser desorption ionisation-time of flight MS. Protein profiling revealed species-specific differences in the proteome, with some protein peaks present in all three species as well as protein peaks that were unique for donkeys and mules, horses and mules or for horses alone. The molecular weight of TTR of horses and donkeys differed by 30Da, and both species revealed several modified forms of TTR besides the native form. The mass spectra of mules represented a merging of TTR spectra of horses and donkeys. In summary, the present study indicated that there are substantial differences in the proteome of horses and donkeys. Additionally, the results probably indicate that the proteome of mules reveal a higher similarity to donkeys than to horses.
The Arctic is a particularly sensitive area with respect to climate change due to the high surface albedo of snow and ice and the extreme radiative conditions. Clouds and aerosols as parts of the Arctic atmosphere play an important role in the radiation budget, which is, as yet, poorly quantified and understood. The LIDAR (Light Detection And Ranging) measurements presented in this PhD thesis contribute with continuous altitude resolved aerosol profiles to the understanding of occurrence and characteristics of aerosol layers above Ny-Ålesund, Spitsbergen. The attention was turned to the analysis of periods with high aerosol load. As the Arctic spring troposphere exhibits maximum aerosol optical depths (AODs) each year, March and April of both the years 2007 and 2009 were analyzed. Furthermore, stratospheric aerosol layers of volcanic origin were analyzed for several months, subsequently to the eruptions of the Kasatochi and Sarychev volcanoes in summer 2008 and 2009, respectively. The Koldewey Aerosol Raman LIDAR (KARL) is an instrument for the active remote sensing of atmospheric parameters using pulsed laser radiation. It is operated at the AWIPEV research base and was fundamentally upgraded within the framework of this PhD project. It is now equipped with a new telescope mirror and new detection optics, which facilitate atmospheric profiling from 450m above sea level up to the mid-stratosphere. KARL provides highly resolved profiles of the scattering characteristics of aerosol and cloud particles (backscattering, extinction and depolarization) as well as water vapor profiles within the lower troposphere. Combination of KARL data with data from other instruments on site, namely radiosondes, sun photometer, Micro Pulse LIDAR, and tethersonde system, resulted in a comprehensive data set of scattering phenomena in the Arctic atmosphere. The two spring periods March and April 2007 and 2009 were at first analyzed based on meteorological parameters, like local temperature and relative humidity profiles as well as large scale pressure patterns and air mass origin regions. Here, it was not possible to find a clear correlation between enhanced AOD and air mass origin. However, in a comparison of two cloud free periods in March 2007 and April 2009, large AOD values in 2009 coincided with air mass transport through the central Arctic. This suggests the occurrence of aerosol transformation processes during the aerosol transport to Ny-Ålesund. Measurements on 4 April 2009 revealed maximum AOD values of up to 0.12 and aerosol size distributions changing with altitude. This and other performed case studies suggest the differentiation between three aerosol event types and their origin: Vertically limited aerosol layers in dry air, highly variable hygroscopic boundary layer aerosols and enhanced aerosol load across wide portions of the troposphere. For the spring period 2007, the available KARL data were statistically analyzed using a characterization scheme, which is based on optical characteristics of the scattering particles. The scheme was validated using several case studies. Volcanic eruptions in the northern hemisphere in August 2008 and June 2009 arose the opportunity to analyze volcanic aerosol layers within the stratosphere. The rate of stratospheric AOD change was similar within both years with maximum values above 0.1 about three to five weeks after the respective eruption. In both years, the stratospheric AOD persisted at higher rates than usual until the measurements were stopped in late September due to technical reasons. In 2008, up to three aerosol layers were detected, the layer structure in 2009 was characterized by up to six distinct and thin layers which smeared out to one broad layer after about two months. The lowermost aerosol layer was continuously detected at the tropopause altitude. Three case studies were performed, all revealed rather large indices of refraction of m = (1.53–1.55) - 0.02i, suggesting the presence of an absorbing carbonaceous component. The particle radius, derived with inversion calculations, was also similar in both years with values ranging from 0.16 to 0.19 μm. However, in 2009, a second mode in the size distribution was detected at about 0.5 μm. The long term measurements with the Koldewey Aerosol Raman LIDAR in Ny-Ålesund provide the opportunity to study Arctic aerosols in the troposphere and the stratosphere not only in case studies but on longer time scales. In this PhD thesis, both, tropospheric aerosols in the Arctic spring and stratospheric aerosols following volcanic eruptions have been described qualitatively and quantitatively. Case studies and comparative studies with data of other instruments on site allowed for the analysis of microphysical aerosol characteristics and their temporal evolution.
On the 20.01.1991 the Latvian people defended the Latvian political elite from the Soviet OMON troops in order to achieve independence. After this impressive sign of civil society the people fell asleep, the level of mobility and the satisfaction with the functioning of democracy therefore is rather weak. The referendum (2008), to gain the right to dissolve the Parliament by the people, initiated by the Trade Unions can be assessed as a sign that there is something on the move. This paper is trying to give an impression of the situation of the civil society in terms of participation in the decision- making process. Hereby the focus lays on NGOs: What is the legal base and which problems do they face. To learn more about the situation interviews were organized with representatives of NGOs from different sectors like community development; Social inclusion; advocating gender issues as well as environment and sustainable development. As a result of the research it can be said that the civil society made some steps forward but it is still struggling with a high level of corruption, lack of interested from the elite and the ordinary people and the insecure financial state.
Active Galactic Nuclei (AGN) are powered by gas accretion onto supermassive Black Holes (BH). The luminosity of AGN can exceed the integrated luminosity of their host galaxies by orders of magnitude, which are then classified as Quasi-Stellar Objects (QSOs). Some mechanisms are needed to trigger the nuclear activity in galaxies and to feed the nuclei with gas. Among several possibilities, such as gravitational interactions, bar instabilities, and smooth gas accretion from the environment, the dominant process has yet to be identified. Feedback from AGN may be important an important ingredient of the evolution of galaxies. However, the details of this coupling between AGN and their host galaxies remain unclear. In this work we aim to investigate the connection between the AGN and their host galaxies by studying the properties of the extendend ionised gas around AGN. Our study is based on observations of ~50 luminous, low-redshift (z<0.3) QSOs using the novel technique of integral field spectroscopy that combines imaging and spectroscopy. After spatially separating the emission of AGN-ionised gas from HII regions, ionised solely by recently formed massive stars, we demonstrate that the specific star formation rates in several disc-dominated AGN hosts are consistent with those of normal star forming galaxies, while others display no detectable star formation activity. Whether the star formation has been actively suppressed in those particular host galaxies by the AGN, or their gas content is intrinsically low, remains an open question. By studying the kinematics of the ionised gas, we find evidence for non-gravitational motions and outflows on kpc scales only in a few objects. The gas kinematics in the majority of objects however indicate a gravitational origin. It suggests that the importance of AGN feedback may have been overrated in theoretical works, at least at low redshifts. The [OIII] line is the strongest optical emission line for AGN-ionised gas, which can be extended over several kpc scales, usually called the Narrow-Line Region (NLR). We perform a systematic investigation of the NLR size and determine a NLR size-luminosity relation that is consistent with the scenario of a constant ionisation parameter throughout the NLR. We show that previous narrow-band imaging with the Hubble Space Telescope underestimated the NLR size by a factor of >2 and that the continuum AGN luminosity is better correlated with the NLR size than the [OIII] luminosity. These affects may account for the different NLR size-luminosity relations reported in previous studies. On the other hand, we do not detect extended NLRs around all QSOs, and demonstrate that the detection of extended NLRs goes along with radio emission. We employ emission line ratios as a diagnostic for the abundance of heavy elements in the gas, i.e. its metallicity, and find that the radial metallicity gradients are always flatter than in inactive disc-dominated galaxies. This can be interpreted as evidence for radial gas flows from the outskirts of these galaxies to the nucleus. Recent or ongoing galaxy interactions are likely responsible for this effect and may turn out to be a common prerequisite for QSO activity. The metallicity of bulge-dominated hosts are systematically lower than their disc-dominated counterparts, which we interpret as evidence for minor mergers, supported by our detailed study of the bulge-dominated host of the luminous QSO HE 1029-1401, or smooth gas accretion from the environment. In this line another new discovery is that HE 2158-0107 at z=0.218 is the most metal poor luminous QSO ever observed. Together with a large (30kpc) extended structure of low metallicity ionised gas, we propose smooth cold gas accretion as the most likely scenario. Theoretical studies suggested that this process is much more important at earlier epochs of the universe, so that HE 2158-0107 might be an ideal laboratory to study this mechanism of galaxy and BH growth at low redshift more detailed in the furture.
Due to their optical and electro-conductive attributes, carbazole derivatives are interesting materials for a large range of biosensor applications. In this study, we present the synthesis routes and fluorescence evaluation of newly designed carbazole fluorosensors that, by modification with uracil, have a special affinity for antiretroviral drugs via either Watson–Crick or Hoogsteen base pairing. To an N-octylcarbazole-uracil compound, four different groups were attached, namely thiophene, furane, ethylenedioxythiophene, and another uracil; yielding four different derivatives. Photophysical properties of these newly obtained derivatives are described, as are their interactions with the reverse transcriptase inhibitors such as abacavir, zidovudine, lamivudine and didanosine. The influence of each analyte on biosensor fluorescence was assessed on the basis of the Stern–Volmer equation and represented by Stern–Volmer constants. Consequently we have demonstrated that these structures based on carbazole, with a uracil group, may be successfully incorporated into alternative carbazole derivatives to form biosensors for the molecular recognition of antiretroviral drugs.
Plant growth and survival depend on photosynthesis in the leaves. This involves the uptake of carbon dioxide from the atmosphere and the simultaneous capture of light energy to produce organic molecules, which enter metabolism and are converted to many other compounds which then serve as building blocks for biomass growth. Leaves are organs specialised for photosynthetic carbon dioxide fixation. The function of leaves involves many trade-offs which must be optimised in order to achieve effective use of resources and maximum photosynthesis. It is known that the morphology of leaves adjusts to the growth environment of plants and this is important for optimising their function for photosynthesis. However, it is unclear how this adjustment is regulated. The general aim of the work presented in this thesis is to understand how leaf growth and morphology are regulated in the model species Arabidopsis thaliana. Special attention was dedicated to the possibility that there might be internal metabolic signals within the plant which affect the growth and development of leaves. In order to investigate this question, leaf growth and development must be considered beyond the level of the single organ and in the context of the whole plant because leaves do not grow autonomously but depend on resources and regulatory influences delivered by the rest of the plant. Due to the complexity of this question, three complementary approaches were taken. In the first and most specific approach it was asked whether a proposed down-stream component of sucrose signalling, trehalose-6-phosphate (Tre-6-P), might influence leaf development and growth. To investigate this question, transgenic Arabidopsis lines with perturbed levels of Tre-6-P were generated using the constitutive 35S promoter to express bacterial enzymes involved in trehalose metabolism. These experiments also led to an unanticipated project concerning a possible role for Tre-6-P in stomatal function, which is another very important function in leaves. In a second and more general approach it was investigated whether changes in sugar levels in plants affect the morphogenesis of leaves in response to light. For this, a series of metabolic mutants impaired in central metabolism were grown in one light environment and their leaf morphology was analysed. In a third and even more general approach the natural variation in leaf and rosette morphological traits was investigated in a panel of wild Arabidopsis accessions with the aim of understanding how leaf morphology affects leaf function and whole plant growth and how different traits relate to each other. The analysis included measurements of leaf morphological traits as well as the number of leaves in the plant to put leaf morphology in a whole plant context. The variance in plant growth could not be explained by variation in photosynthetic rates and only to a small degree by variation in rates of dark respiration. There were four key axes of variation in rosette and leaf morphology – leaf area growth, leaf thickness, cell expansion and leaf number. These four processes were integrated in the context of whole plant growth by models that employed a multiple linear regression approach. This then led to a theoretical approach in which a simple allometric mathematical model was constructed, linking leaf number, leaf size and plant growth rate together in a whole plant context in Arabidopsis.
Motivation | Societal and economic needs of East Africa rely entirely on the availability of water, which is governed by the regular onset and retreat of the rainy seasons. Fluctuations in the amounts of rainfall has tremendous impact causing widespread famine, disease outbreaks and human migrations. Efforts towards high resolution forecasting of seasonal precipitation and hydrological systems are therefore needed, which requires high frequency short to long-term analyses of available climate data that I am going to present in this doctoral thesis by three different studies. 15,000 years - Suguta Valley | The main study of this thesis concentrated on the understanding of humidity changes within the last African Humid Period (AHP, 14.8-5.5 ka BP). The nature and causes of intensity variations of the West-African (WAM) and Indian Summer monsoons (ISM) during the AHP, especially their exact influence on regional climate relative to each other, is currently intensely debated. Here, I present a high-resolution multiproxy lake-level record spanning the AHP from the remote Suguta Valley in the northern Kenya Rift, located between the WAM and ISM domains. The presently desiccated valley was during the AHP filled by a 300 m deep and 2200 km2 large palaeo-lake due to an increase in precipitation of only 26%. The record explains the synchronous onset of large lakes in the East African Rift System (EARS) with the longitudinal shift of the Congo Air Boundary (CAB) over the East African and Ethiopian Plateaus, as the direct consequence of an enhanced atmospheric pressure gradient between East-Africa and India due to a precessional-forced northern hemisphere insolation maximum. Pronounced, and abrupt lake level fluctuations during the generally wet AHP are explained by small-scale solar irradiation changes weakening this pressure gradient atmospheric moisture availability preventing the CAB from reaching the study area. Instead, the termination of the AHP occurred, in a non-linear manner due to a change towards an equatorial insolation maximum ca. 6.5 ka ago extending the AHP over Ethiopia and West-Africa. 200 years - Lake Naivasha | The second part of the thesis focused on the analysis of a 200 year-old sediment core from Lake Naivasha in the Central Kenya Rift, one of the very few present freshwater lakes in East Africa. The results revealed and confirmed, that the appliance of proxy records for palaeo-climate reconstruction for the last 100 years within a time of increasing industrialisation and therefore human impact to the proxy-record containing sites are broadly limited. Since the middle of the 20th century, intense anthropogenic activity around Lake Naivasha has led to cultural eutrophication, which has overprinted the influence of natural climate variation to the lake usually inferred from proxy records such as diatoms, transfer-functions, geochemical and sedimentological analysis as used in this study. The results clarify the need for proxy records from remote unsettled areas to contribute with pristine data sets to current debates about anthropologic induced global warming since the past 100 years. 14 years - East African Rift | In order to avoid human influenced data sets and validate spatial and temporal heterogeneities of proxy-records from East Africa, the third part of the thesis therefore concentrated on the most recent past 14 years (1996-2010) detecting climate variability by using remotely sensed rainfall data. The advancement in the spatial coverage and temporal resolutions of rainfall data allow a better understanding of influencing climate mechanisms and help to better interpret proxy-records from the EARS in order to reconstruct past climate conditions. The study focuses on the dynamics of intraseasonal rainfall distribution within catchments of eleven lake basins in the EARS that are often used for palaeo-climate studies. We discovered that rainfall in adjacent basins exhibits high complexities in the magnitudes of intraseasonal variability, biennial to triennial precipitation patterns and even are not necessarily correlated often showing opposite trends. The variability among the watersheds is driven by the complex interaction of topography, in particular the shape, length and elevation of the catchment and its relative location to the East African Rift System and predominant influence of the ITCZ or CAB, whose locations and intensities are dependent on the strength of low pressure cells over India, SST variations in the Atlantic, Pacific or Indian Ocean, QBO phases and the 11-year solar cycle. Among all seasons we observed, January-September is the season of highest and most complex rainfall variability, especially for the East African Plateau basins, most likely due to the irregular penetration and sensitivity of the CAB.
This work addresses issues in the automatic preprocessing of historical German input text for use by conventional natural language processing techniques. Conventional techniques cannot adequately account for historical input text due to conventional tools' reliance on a fixed application-specific lexicon keyed by contemporary orthographic surface form on the one hand, and the lack of consistent orthographic conventions in historical input text on the other. Historical spelling variation is treated here as an error-correction problem or "canonicalization" task: an attempt to automatically assign each (historical) input word a unique extant canonical cognate, thus allowing direct application-specific processing (tagging, parsing, etc.) of the returned canonical forms without need for any additional application-specific modifications. In the course of the work, various methods for automatic canonicalization are investigated and empirically evaluated, including conflation by phonetic identity, conflation by lemma instantiation heuristics, canonicalization by weighted finite-state rewrite cascade, and token-wise disambiguation by a dynamic Hidden Markov Model.
Organic thin film transistors (TFT) are an attractive option for low cost electronic applications and may be used for active matrix displays and for RFID applications. To extend the range of applications there is a need to develop and optimise the performance of non-volatile memory devices that are compatible with the solution-processing fabrication procedures used in plastic electronics. A possible candidate is an organic TFT incorporating the ferroelectric co-polymer poly(vinylidenefluoride-trifluoroethylene)(P(VDF-TrFE)) as the gate insulator. Dielectric measurements have been carried out on all-organic metal-insulator-semiconductor structures with the ferroelectric polymer poly(vinylidenefluoride-trifluoroethylene) (P(VDF-TrFE)) as the gate insu-lator. The capacitance spectra of MIS devices, were measured under different biases, showing the effect of charge accumulation and depletion on the Maxwell-Wagner peak. The position and height of this peak clearly indicates the lack of stable depletion behavior and the decrease of mobility when increasing the depletion zone width, i.e. upon moving into the P3HT bulk. The lack of stable depletion was further investigated with capacitance-voltage (C-V) measurements. When the structure was driven into depletion, C-V plots showed a positive flat-band voltage shift, arising from the change in polarization state of the ferroelectric insulator. When biased into accumulation, the polarization was reversed. It is shown that the two polarization states are stable i.e. no depolarization occurs below the coercive field. However, negative charge trapped at the semiconductor-insulator interface during the depletion cycle masks the negative shift in flat-band voltage expected during the sweep to accumulation voltages. The measured output characteristics of the studied ferroelectric-field-effect transistors confirmed the results of the C-V plots. Furthermore, the results indicated a trapping of electrons at the positively charged surfaces of the ferroelectrically polarized P(VDF-TrFE) crystallites near the insulator/semiconductor in-terface during the first poling cycles. The study of the MIS structure by means of thermally stimulated current (TSC) revealed further evidence for the stability of the polarization under depletion voltages. It was shown, that the lack of stable depletion behavior is caused by the compensation of the orientational polarization by fixed electrons at the interface and not by the depolarization of the insulator, as proposed in several publications. The above results suggest a performance improvement of non-volatile memory devices by the optimization of the interface.
Plants and some unicellular algae store carbon in the form of transitory starch on a diurnal basis. The turnover of this glucose polymer is tightly regulated and timely synthesis as well as mobilization is essential to provide energy for heterotrophic growth. Especially for starch degradation, novel enzymes and mechanisms have been proposed recently. However, the catalytic properties of these enzymes and their coordination with metabolic regulation are still to be discovered. This thesis develops theoretical methods in order to interpret and analyze enzymes and their role in starch degradation. In the first part, a novel description of interfacial enzyme catalysis is proposed. Since the initial steps of starch degradation involve reactions at the starch-stroma interface it is necessary to have a framework which allows the derivation of interfacial enzyme rate laws. A cornerstone of the method is the introduction of the available area function - a concept from surface physics - to describe the adsorption step in the catalytic cycle. The method is applied to derive rate laws for two hydrolases, the Beta-amylase (BAM3) and the Isoamylase (DBE/ISA3), as well as to the Glucan, water dikinase (GWD) and a Phosphoglucan phosphatase (DSP/SEX4). The second part uses the interfacial rate laws to formulate a kinetic model of starch degradation. It aims at reproducing the stimulatory effect of reversible phosphorylation by GWD and DSP on the breakdown of the granule. The model can describe the dynamics of interfacial properties during degradation and suggests that interfacial amylopectin side-chains undergo spontaneous helix-coil transitions. Reversible phosphorylation has a synergistic effect on glucan release especially in the early phase dropping off during degradation. Based on the model, the hypothesis is formulated that interfacial phosphorylation is important for the rapid switch from starch synthesis to starch degradation. The third part takes a broader perspective on carbohydrate-active enzymes (CAZymes) but is motivated by the organization of the downstream pathway of starch breakdown. This comprises Alpha-1,4-glucanotransferases (DPE1 and DPE2) and Alpha-glucan-phosphorylases (Pho or PHS) both in the stroma and in the cytosol. CAZymes accept many different substrates and catalyze numerous reactions and therefore cannot be characterized in classical enzymological terms. A concise characterization is provided by conceptually linking statistical thermodynamics and polymer biochemistry. Each reactant is interpreted as an energy level, transitions between which are constrained by the enzymatic mechanisms. Combinations of in vitro assays of polymer-active CAZymes essential for carbon metabolism in plants confirmed the dominance of entropic gradients. The principle of entropy maximization provides a generalization of the equilibrium constant. Stochastic simulations confirm the results and suggest that randomization of metabolites in the cytosolic pool of soluble heteroglycans (SHG) may contribute to a robust integration of fluctuating carbon fluxes coming from chloroplasts.
This Thesis puts its focus on the physics of neutron stars and its description with methods of numerical relativity. In the first step, a new numerical framework the Whisky2D code will be developed, which solves the relativistic equations of hydrodynamics in axisymmetry. Therefore we consider an improved formulation of the conserved form of these equations. The second part will use the new code to investigate the critical behaviour of two colliding neutron stars. Considering the analogy to phase transitions in statistical physics, we will investigate the evolution of the entropy of the neutron stars during the whole process. A better understanding of the evolution of thermodynamical quantities, like the entropy in critical process, should provide deeper understanding of thermodynamics in relativity. More specifically, we have written the Whisky2D code, which solves the general-relativistic hydrodynamics equations in a flux-conservative form and in cylindrical coordinates. This of course brings in 1/r singular terms, where r is the radial cylindrical coordinate, which must be dealt with appropriately. In the above-referenced works, the flux operator is expanded and the 1/r terms, not containing derivatives, are moved to the right-hand-side of the equation (the source term), so that the left hand side assumes a form identical to the one of the three-dimensional (3D) Cartesian formulation. We call this the standard formulation. Another possibility is not to split the flux operator and to redefine the conserved variables, via a multiplication by r. We call this the new formulation. The new equations are solved with the same methods as in the Cartesian case. From a mathematical point of view, one would not expect differences between the two ways of writing the differential operator, but, of course, a difference is present at the numerical level. Our tests show that the new formulation yields results with a global truncation error which is one or more orders of magnitude smaller than those of alternative and commonly used formulations. The second part of the Thesis uses the new code for investigations of critical phenomena in general relativity. In particular, we consider the head-on-collision of two neutron stars in a region of the parameter space where two final states a new stable neutron star or a black hole, lay close to each other. In 1993, Choptuik considered one-parameter families of solutions, S[P], of the Einstein-Klein-Gordon equations for a massless scalar field in spherical symmetry, such that for every P > P⋆, S[P] contains a black hole and for every P < P⋆, S[P] is a solution not containing singularities. He studied numerically the behavior of S[P] as P → P⋆ and found that the critical solution, S[P⋆], is universal, in the sense that it is approached by all nearly-critical solutions regardless of the particular family of initial data considered. All these phenomena have the common property that, as P approaches P⋆, S[P] approaches a universal solution S[P⋆] and that all the physical quantities of S[P] depend only on |P − P⋆|. The first study of critical phenomena concerning the head-on collision of NSs was carried out by Jin and Suen in 2007. In particular, they considered a series of families of equal-mass NSs, modeled with an ideal-gas EOS, boosted towards each other and varied the mass of the stars, their separation, velocity and the polytropic index in the EOS. In this way they could observe a critical phenomenon of type I near the threshold of black-hole formation, with the putative solution being a nonlinearly oscillating star. In a successive work, they performed similar simulations but considering the head-on collision of Gaussian distributions of matter. Also in this case they found the appearance of type-I critical behaviour, but also performed a perturbative analysis of the initial distributions of matter and of the merged object. Because of the considerable difference found in the eigenfrequencies in the two cases, they concluded that the critical solution does not represent a system near equilibrium and in particular not a perturbed Tolmann-Oppenheimer-Volkoff (TOV) solution. In this Thesis we study the dynamics of the head-on collision of two equal-mass NSs using a setup which is as similar as possible to the one considered above. While we confirm that the merged object exhibits a type-I critical behaviour, we also argue against the conclusion that the critical solution cannot be described in terms of equilibrium solution. Indeed, we show that, in analogy with what is found in, the critical solution is effectively a perturbed unstable solution of the TOV equations. Our analysis also considers fine-structure of the scaling relation of type-I critical phenomena and we show that it exhibits oscillations in a similar way to the one studied in the context of scalar-field critical collapse.
Aggregation of the Amyloid β (Aβ) peptide to amyloid fibrils is associated with the outbreak of Alzheimer’s disease. Early aggregation intermediates in form of soluble oligomers are of special interest as they are believed to be the major toxic components in the process. These oligomers are of disordered and transient nature. Therefore, their detailed molecular structure is difficult to access experimentally and often remains unknown. In the present work extensive, fully atomistic replica exchange molecular dynamics simulations were performed to study the preaggregated, monomer states and early aggregation intermediates (dimers, trimers) of Aβ(25-35) and Aβ(10-35)-NH2 in aqueous solution. The folding and aggregation of Aβ(25-35) were studied at neutral pH and 293 K. Aβ(25-35) monomers mainly adopt β-hairpin conformations characterized by a β-turn formed by residues G29 and A30, and a β-sheet between residues N27–K28 and I31–I32 in equilibrium with coiled conformations. The β-hairpin conformations served as initial configurations to model spontaneous aggregation of Aβ(25-35). As expected, within the Aβ(25-35) dimer and trimer ensembles many different poorly populated conformations appear. Nevertheless, we were able to distinguish between disordered and fibril-like oligomers. Whereas disordered oligomers are rather compact with few intermolecular hydrogen bonds (HBs), fibril-like oligomers are characterized by the formation of large intermolecular β-sheets. In most of the fibril-like dimers and trimers individual peptides are fully extended forming in- or out-of-register antiparallel β-sheets. A small amount of fibril-like trimers contained V-shaped peptides forming parallel β-sheets. The dimensions of extended and V-shaped oligomers correspond well to the diameters of two distinct morphologies found for Aβ(25-35) fibrils. The transition from disordered to fibril-like Aβ(25-35) dimers is unfavorable but driven by energy. The lower energy of fibril-like dimers arises from favorable intermolecular HBs and other electrostatic interactions which compete with a loss in entropy. Approximately 25 % of the entropic cost correspond to configurational entropy. The rest relates to solvent entropy, presumably caused by hydrophobic and electrostatic effects. In contrast to the transition towards fibril-like dimers the first step of aggregation is driven by entropy. Here, we compared structural and thermodynamic properties of the individual monomer, dimer and trimer ensembles to gain qualitative information about the aggregation process. The β-hairpin conformation observed for monomers is successively dissolved in dimer and trimer ensembles while instead intermolecular β-sheets are formed. As expected upon aggregation the configurational entropy decreases. Additionally, the solvent accessible surface area (SASA), especially the hydrophobic SASA, decreases yielding a favorable solvation free energy which overcompensates the loss in configurational entropy. In summary, the hydrophobic effect, possibly combined with electrostatic effects, yields an increase in solvent entropy which is believed to be one major driving force towards aggregation. Spontaneous folding of the Aβ(10-35)-NH2 monomer was modeled using two force fields, GROMOS96 43a1 and OPLS/AA, and compared to primary NMR data collected at pH 5.6 and 283 K taken from the literature. Unexpectedly, the two force fields yielded significantly different main conformations. Comparison between experimental and calculated nuclear Overhauser effect (NOE) distances is not sufficient to distinguish between the different force fields. Additionally, the comparison with scalar coupling constants suggest that the chosen protonation in both simulations corresponds to a pH lower than in the experiment. Based on this analysis we were unable to determine which force field yields a better description of this system. Dimerization of Aβ(10-35)-NH2 was studied at neutral pH and 300 K. Dimer conformations arrange in many distinct, poorly populated and rather complex alignments or interlocking patterns which are rather stabilized by side chain interactions than by specific intermolecular hydrogen bonds. Similar to Aβ(25-35) dimers, transition towards β-sheet-rich, fibril-like Aβ(10-35) dimers is driven by energy competing with a loss in entropy. Here, transition is mediated by favorable peptide-solvent and solvent-solvent interactions mainly arising from electrostatic interactions.
Version Control Systems (VCS) allow developers to manage changes to software artifacts. Developers interact with VCSs through a variety of client programs, such as graphical front-ends or command line tools. It is desirable to use the same version control client program against different VCSs. Unfortunately, no established abstraction over VCS concepts exists. Instead, VCS client programs implement ad-hoc solutions to support interaction with multiple VCSs. This thesis presents Pur, an abstraction over version control concepts that allows building rich client programs that can interact with multiple VCSs. We provide an implementation of this abstraction and validate it by implementing a client application.
There has been a substantial increase in the percentage for publications with co-authors located in departments from different countries in 12 major journals of psychology. The results are evidence for a remarkable internationalization of psychological research, starting in the mid 1970s and increasing in rate at the beginning of the 1990s. This growth occurs against a constant number of articles with authors from the same country; it is not due to a concomitant increase in the number of co-authors per article. Thus, international collaboration in psychology is obviously on the rise.
Linear mixed models (LMMs) provide a still underused methodological perspective on combining experimental and individual-differences research. Here we illustrate this approach with two-rectangle cueing in visual attention (Egly et al., 1994). We replicated previous experimental cue-validity effects relating to a spatial shift of attention within an object (spatial effect), to attention switch between objects (object effect), and to the attraction of attention toward the display centroid (attraction effect), also taking into account the design-inherent imbalance of valid and other trials. We simultaneously estimated variance/covariance components of subject-related random effects for these spatial, object, and attraction effects in addition to their mean reaction times (RTs). The spatial effect showed a strong positive correlation with mean RT and a strong negative correlation with the attraction effect. The analysis of individual differences suggests that slow subjects engage attention more strongly at the cued location than fast subjects. We compare this joint LMM analysis of experimental effects and associated subject-related variances and correlations with two frequently used alternative statistical procedures
In this thesis, simulations of laser-driven many-electron dynamics in molecules are presented, i.e., the interaction between molecules and an electromagnetic field is demonstrated. When a laser field is applied to a molecular system, a population of higher electronic states takes place as well as other processes, e.g. photoionization, which is described by an appropriate model. Also, a finite lifetime of an excited state can be described by such a model. In the second part, a method is postulated that is capable of describing electron correlation in a time-dependent scheme. This is done by introducing a single-electron entropy that is at least temporarily minimized in a further step.
The modeling and evaluation calculus FMC-QE, the Fundamental Modeling Concepts for Quanti-tative Evaluation [1], extends the Fundamental Modeling Concepts (FMC) for performance modeling and prediction. In this new methodology, the hierarchical service requests are in the main focus, because they are the origin of every service provisioning process. Similar to physics, these service requests are a tuple of value and unit, which enables hierarchical service request transformations at the hierarchical borders and therefore the hierarchical modeling. Through reducing the model complexity of the models by decomposing the system in different hierarchical views, the distinction between operational and control states and the calculation of the performance values on the assumption of the steady state, FMC-QE has a scalable applica-bility on complex systems. According to FMC, the system is modeled in a 3-dimensional hierarchical representation space, where system performance parameters are described in three arbitrarily fine-grained hierarchi-cal bipartite diagrams. The hierarchical service request structures are modeled in Entity Relationship Diagrams. The static server structures, divided into logical and real servers, are de-scribed as Block Diagrams. The dynamic behavior and the control structures are specified as Petri Nets, more precisely Colored Time Augmented Petri Nets. From the structures and pa-rameters of the performance model, a hierarchical set of equations is derived. The calculation of the performance values is done on the assumption of stationary processes and is based on fundamental laws of the performance analysis: Little's Law and the Forced Traffic Flow Law. Little's Law is used within the different hierarchical levels (horizontal) and the Forced Traffic Flow Law is the key to the dependencies among the hierarchical levels (vertical). This calculation is suitable for complex models and allows a fast (re-)calculation of different performance scenarios in order to support development and configuration decisions. Within the Research Group Zorn at the Hasso Plattner Institute, the work is embedded in a broader research in the development of FMC-QE. While this work is concentrated on the theoretical background, description and definition of the methodology as well as the extension and validation of the applicability, other topics are in the development of an FMC-QE modeling and evaluation tool and the usage of FMC-QE in the design of an adaptive transport layer in order to fulfill Quality of Service and Service Level Agreements in volatile service based environments. This thesis contains a state-of-the-art, the description of FMC-QE as well as extensions of FMC-QE in representative general models and case studies. In the state-of-the-art part of the thesis in chapter 2, an overview on existing Queueing Theory and Time Augmented Petri Net models and other quantitative modeling and evaluation languages and methodologies is given. Also other hierarchical quantitative modeling frameworks will be considered. The description of FMC-QE in chapter 3 consists of a summary of the foundations of FMC-QE, basic definitions, the graphical notations, the FMC-QE Calculus and the modeling of open queueing networks as an introductory example. The extensions of FMC-QE in chapter 4 consist of the integration of the summation method in order to support the handling of closed networks and the modeling of multiclass and semaphore scenarios. Furthermore, FMC-QE is compared to other performance modeling and evaluation approaches. In the case study part in chapter 5, proof-of-concept examples, like the modeling of a service based search portal, a service based SAP NetWeaver application and the Axis2 Web service framework will be provided. Finally, conclusions are given by a summary of contributions and an outlook on future work in chapter 6. [1] Werner Zorn. FMC-QE - A New Approach in Quantitative Modeling. In Hamid R. Arabnia, editor, Procee-dings of the International Conference on Modeling, Simulation and Visualization Methods (MSV 2007) within WorldComp ’07, pages 280 – 287, Las Vegas, NV, USA, June 2007. CSREA Press. ISBN 1-60132-029-9.
In this thesis chemical reactions under hydrothermal conditions were explored, whereby emphasis was put on green chemistry. Water at high temperature and pressure acts as a benign solvent. Motivation to work under hydrothermal conditions was well-founded in the tunability of physicochemical properties with temperature, e.g. of dielectric constant, density or ion product, which often resulted in surprising reactivity. Another cornerstone was the implementation of the principles of green chemistry. Besides the use of water as solvent, this included the employment of a sustainable feedstock and the sensible use of resources by minimizing waste and harmful intermediates and additives. To evaluate the feasibility of hydrothermal conditions for chemical synthesis, exemplary reactions were performed. These were carried out in a continuous flow reactor, allowing for precise control of reaction conditions and kinetics measurements. In most experiments a temperature of 200 °C in combination with a pressure of 100 bar was chosen. In some cases the temperature was even raised to 300 °C. Water in this subcritical range can also be found in nature at hydrothermal vents on the ocean floor. On the primitive earth, environments with such conditions were however present in larger numbers. Therefore we tested whether biologically important carbohydrates could be formed at high temperature from the simple, probably prebiotic precursor formaldehyde. Indeed, this formose reaction could be carried out successfully, although the yield was lower compared to the counterpart reaction under ambient conditions. However, striking differences regarding selectivity and necessary catalysts were observed. At moderate temperatures bases and catalytically active cations like Ca2+ are necessary and the main products are hexoses and pentoses, which accumulate due to their higher stability. In contrast, in high-temperature water no catalyst was necessary but a slightly alkaline solution was sufficient. Hexoses were only formed in negligible amounts, whereas pentoses and the shorter carbohydrates accounted for the major fraction. Amongst the pentoses there was some preference for the formation of ribose. Even deoxy sugars could be detected in traces. The observation that catalysts can be avoided was successfully transferred to another reaction. In a green chemistry approach platform chemicals must be produced from sustainable resources. Carbohydrates can for instance be employed as a basis. They can be transformed to levulinic acid and formic acid, which can both react via a transfer hydrogenation to the green solvent and biofuel gamma-valerolactone. This second reaction usually requires catalysis by Ru or Pd, which are neither sustainable nor low-priced. Under hydrothermal conditions these heavy metals could be avoided and replaced by cheap salts, taking advantage of the temperature dependence of the acid dissociation constant. Simple sulfate was recognized as a temperature switchable base. With this additive high yield could be achieved by simultaneous prevention of waste. In contrast to conventional bases, which create salt upon neutralization, a temperature switchable base becomes neutral again when cooled down and thus can be reused. This adds another sustainable feature to the high atom economy of the presented hydrothermal synthesis. In a last study complex decomposition pathways of biomass were investigated. Gas chromatography in conjunction with mass spectroscopy has proven to be a powerful tool for the identification of unknowns. It was observed that several acids were formed when carbohydrates were treated with bases at high temperature. This procedure was also applied to digest wood. Afterwards it was possible to fermentate the solution and a good yield of methane was obtained. This has to be regarded in the light of the fact that wood practically cannot be used as a feedstock in a biogas factory. Thus the hydrothermal pretreatment is an efficient means to employ such materials as well. Also the reaction network of the hydrothermal decomposition of glycine was investigated using isotope-labeled compounds as comparison for the unambiguous identification of unknowns. This refined analysis allowed the identification of several new molecules and pathways, not yet described in literature. In summary several advantages could be taken from synthesis in high-temperature water. Many catalysts, absolutely necessary under ambient conditions, could either be completely avoided or replaced by cheap, sustainable alternatives. In this respect water is not only a green solvent, but helps to prevent waste and preserves resources.
The present thesis introduces an iterative expert-based Bayesian approach for assessing greenhouse gas (GHG) emissions from the 2030 German new vehicle fleet and quantifying the impacts of their main drivers. A first set of expert interviews has been carried out in order to identify technologies which may help to lower car GHG emissions and to quantify their emission reduction potentials. Moreover, experts were asked for their probability assessments that the different technologies will be widely adopted, as well as for important prerequisites that could foster or hamper their adoption. Drawing on the results of these expert interviews, a Bayesian Belief Network has been built which explicitly models three vehicle types: Internal Combustion Engine Vehicles (which include mild and full Hybrid Electric Vehicles), Plug-In Hybrid Electric Vehicles, and Battery Electric Vehicles. The conditional dependencies of twelve central variables within the BBN - battery energy, fuel and electricity consumption, relative costs, and sales shares of the vehicle types - have been quantified by experts from German car manufacturers in a second series of interviews. For each of the seven second-round interviews, an expert's individually specified BBN results. The BBN have been run for different hypothetical 2030 scenarios which differ, e.g., in regard to battery development, regulation, and fuel and electricity GHG intensities. The present thesis delivers results both in regard to the subject of the investigation and in regard to its method. On the subject level, it has been found that the different experts expect 2030 German new car fleet emission to be at 50 to 65% of 2008 new fleet emissions under the baseline scenario. They can be further reduced to 40 to 50% of the emissions of the 2008 fleet though a combination of a higher share of renewables in the electricity mix, a larger share of biofuels in the fuel mix, and a stricter regulation of car CO$_2$ emissions in the European Union. Technically, 2030 German new car fleet GHG emissions can be reduced to a minimum of 18 to 44% of 2008 emissions, a development which can not be triggered by any combination of measures modeled in the BBN alone but needs further commitment. Out of a wealth of existing BBN, few have been specified by individual experts through elicitation, and to my knowledge, none of them has been employed for analyzing perspectives for the future. On the level of methods, this work shows that expert-based BBN are a valuable tool for making experts' expectations for the future explicit and amenable to the analysis of different hypothetical scenarios. BBN can also be employed for quantifying the impacts of main drivers. They have been demonstrated to be a valuable tool for iterative stakeholder-based science approaches.
In this work new fluorinated and non-fluorinated mono- and bifunctional trithiocarbonates of the structure Z-C(=S)-S-R and Z-C(=S)-S-R-S-C(=S)-Z were synthesized for the use as chain transfer agents (CTAs) in the RAFT-process. All newly synthesized CTAs were tested for their efficiency to moderate the free radical polymerization process by polymerizing styrene (M3). Besides characterization of the homopolymers by GPC measurements, end- group analysis of the synthesized block copolymers via 1H-, 19F-NMR, and in some cases also UV-vis spectroscopy, were performed attaching suitable fluorinated moieties to the Z- and/or R-groups of the CTAs. Symmetric triblock copolymers of type BAB and non-symmetric fluorine end- capped polymers were accessible using the RAFT process in just two or one polymerization step. In particular, the RAFT-process enabled the controlled polymerization of hydrophilic monomers such as N-isopropylacrylamide (NIPAM) (M1) as well as N-acryloylpyrrolidine (NAP) (M2) for the A-blocks and of the hydrophobic monomers styrene (M3), 2-fluorostyrene (M4), 3-fluorostyrene (M5), 4-fluorostyrene (M6) and 2,3,4,5,6-pentafluorostyrene (M7) for the B-blocks. The properties of the BAB-triblock copolymers were investigated in dilute, concentrated and highly concentrated aqueous solutions using DLS, turbidimetry, 1H- and 19F-NMR, rheology, determination of the CMC, foam height- and surface tension measurements and microscopy. Furthermore, their ability to stabilize emulsions and microemulsions and the wetting behaviour of their aqueous solutions on different substrates was investigated. The behaviour of the fluorine end-functionalized polymers to form micelles was studied applying DLS measurements in diluted organic solution. All investigated BAB-triblock copolymers were able to form micelles and show surface activity at room temperature in dilute aqueous solution. The aqueous solutions displayed moderate foam formation. With different types and concentrations of oils, the formation of emulsions could be detected using a light microscope. A boosting effect in microemulsions could not be found adding BAB-triblock copolymers. At elevated polymer concentrations, the formation of hydrogels was proved applying rheology measurements.
Nanoporous carbon materials are widely used in industry as adsorbents or catalyst supports, whilst becoming increasingly critical to the developing fields of energy storage / generation or separation technologies. In this thesis, the combined use of carbohydrate hydrothermal carbonisation (HTC) and templating strategies is demonstrated as an efficient route to nanostructured carbonaceous materials. HTC is an aqueous-phase, low-temperature (e.g. 130 – 200 °C) carbonisation, which proceeds via dehydration / poly-condensation of carbon precursors (e.g. carbohydrates and their derivatives), allowing facile access to highly functional carbonaceous materials. Whilst possessing utile, modifiable surface functional groups (e.g. -OH and -C=O-containing moieties), materials synthesised via HTC typically present limited accessible surface area or pore volume. Therefore, this thesis focuses on the development of fabrication routes to HTC materials which present enhanced textural properties and well-defined porosity. In the first discussed synthesis, a combined hard templating / HTC route was investigated using a range of sacrificial inorganic templates (e.g. mesoporous silica beads and macroporous alumina membranes (AAO)). Via pore impregnation of mesoporous silica beads with a biomass-derived carbon source (e.g. 2-furaldehyde) and subsequent HTC at 180 oC, an inorganic / carbonaceous hybrid material was produced. Removal of the template component by acid etching revealed the replication of the silica into mesoporous carbonaceous spheres (particle size ~ 5 μm), representing the inverse morphological structure of the original inorganic body. Surface analysis (e.g. FTIR) indicated a material decorated with hydrophilic (oxygenated) functional groups. Further thermal treatment at increasingly elevated temperatures (e.g. at 350, 550, 750 oC) under inert atmosphere allowed manipulation of functionalities from polar hydrophilic to increasingly non-polar / hydrophobic structural motifs (e.g. extension of the aromatic / pseudo-graphitic nature), thus demonstrating a process capable of simultaneous control of nanostructure and surface / bulk chemistry. As an extension of this approach, carbonaceous tubular nanostructures with controlled surface functionality were synthesised by the nanocasting of uniform, linear macropores of an AAO template (~ 200 nm). In this example, material porosity could be controlled, showing increasingly microporous tube wall features as post carbonisation temperature increased. Additionally, by taking advantage of modifiable surface groups, the introduction of useful polymeric moieties (i.e. grafting of thermoresponsive poly(N-isopropylacrylamide)) was also demonstrated, potentially enabling application of these interesting tubular structures in the fields of biotechnology (e.g. enzyme immobilization) and medicine (e.g. as drug micro-containers). Complimentary to these hard templating routes, a combined HTC / soft templating route for the direct synthesis of ordered porous carbonaceous materials was also developed. After selection of structural directing agents and optimisation of synthesis composition, the F127 triblock copolymer (i.e. ethylene oxide (EO)106 propylene oxide (PO)70 ethylene oxide (EO)106) / D-Fructose system was extensively studied. D-Fructose was found to be a useful carbon precursor as the HTC process could be performed at 130 oC, thus allowing access to stable micellular phase. Thermolytic template removal from the synthesised ordered copolymer / carbon composite yielded functional cuboctahedron single crystalline-like particles (~ 5 μm) with well ordered pore structure of a near perfect cubic Im3m symmetry. N2 sorption analysis revealed a predominantly microporous carbonaceous material (i.e. Type I isotherm, SBET = 257 m2g-1, 79 % microporosity) possessing a pore size of ca. 0.9 nm. The addition of a simple pore swelling additive (e.g. trimethylbenzene (TMB)) to this system was found to direct pore size into the mesopore size domain (i.e. Type IV isotherm, SBET = 116 m2g-1, 60 % mesoporosity) generating pore size of ca. 4 nm. It is proposed that in both cases as HTC proceeds to generate a polyfuran-like network, the organised block copolymer micellular phase is essentially “templated”, either via hydrogen bonding between hydrophilic poly(EO) moiety and the carbohydrate or via hydrophobic interaction between hydrophobic poly(PO) moiety and forming polyfuran-like network, whilst the additive TMB presumably interact with poly(PO) moieties, thus swelling the hydrophobic region expanding the micelle template size further into the mesopore range.
"Unavoidably side by side"
(2011)
Background
More than in other domains the heterogeneous services world in bioinformatics demands for a methodology to classify and relate resources in a both human and machine accessible manner. The Semantic Web, which is meant to address exactly this challenge, is currently one of the most ambitious projects in computer science. Collective efforts within the community have already led to a basis of standards for semantic service descriptions and meta-information. In combination with process synthesis and planning methods, such knowledge about types and services can facilitate the automatic composition of workflows for particular research questions.
Results
In this study we apply the synthesis methodology that is available in the Bio-jETI workflow management framework for the semantics-based composition of EMBOSS services. EMBOSS (European Molecular Biology Open Software Suite) is a collection of 350 tools (March 2010) for various sequence analysis tasks, and thus a rich source of services and types that imply comprehensive domain models for planning and synthesis approaches. We use and compare two different setups of our EMBOSS synthesis domain: 1) a manually defined domain setup where an intuitive, high-level, semantically meaningful nomenclature is applied to describe the input/output behavior of the single EMBOSS tools and their classifications, and 2) a domain setup where this information has been automatically derived from the EMBOSS Ajax Command Definition (ACD) files and the EMBRACE Data and Methods ontology (EDAM). Our experiments demonstrate that these domain models in combination with our synthesis methodology greatly simplify working with the large, heterogeneous, and hence manually intractable EMBOSS collection. However, they also show that with the information that can be derived from the (current) ACD files and EDAM ontology alone, some essential connections between services can not be recognized.
Conclusions
Our results show that adequate domain modeling requires to incorporate as much domain knowledge as possible, far beyond the mere technical aspects of the different types and services. Finding or defining semantically appropriate service and type descriptions is a difficult task, but the bioinformatics community appears to be on the right track towards a Life Science Semantic Web, which will eventually allow automatic service composition methods to unfold their full potential.
The heterogeneity in species assemblages of epigeal spiders was studied in a natural forest and in a managed forest. Additionally the effects of small-scale microhabitat heterogeneity of managed and unmanaged forests were determined by analysing the spider assemblages of three different microhabitat structures (i. vegetation, ii. dead wood. iii. litter cover). The spider were collected in a block design by pitfall traps (n=72) in a 4-week interval. To reveal key environmental factors affecting the spider distribution abiotic and biotic habitat parameters (e.g. vegetation parameters, climate parameters, soil moisture) were assessed around each pitfall trap. A TWINSPAN analyses separated pitfall traps from the natural forest from traps of the managed forest. A subsequent discriminant analyses revealed that the temperature, the visible sky, the plant diversity and the mean diameter at breast height as key discriminant factors between the microhabitat groupings designated by the TWINSPAN analyses. Finally a Redundant analysis (RDA) was done revealing similar environmental factors responsible for the spider species distribution, as a good separation of the different forest types as well as the separation of the microhabitat groupings from the TWINSPAN. Overall the study revealed that the spider communities differed between the forest types as well as between the microhabitat structures and thus species distribution changed within a forest stand on a fine spatial scale. It was documented that the structure of managed forests affects the composition of spider assemblages compared to natural forests significantly and even small scale-heterogeneity seems to influence the spider species composition.
Sediment records of three European lakes were investigated in order to reconstruct the regional climate development during the Lateglacial and Holocene, to investigate the response of local ecosystems to climatic fluctuations and human impact and to relate regional peculiarities of past climate development to climatic changes on a larger spatial scale. The Lake Hańcza (NE Poland) sediment record was studied with a focus on reconstructing the early Holocene climate development and identifying possible differences to Western Europe. Following the initial Holocene climatic improvement, a further climatic improvement occurred between 10 000 and 9000 cal. a BP. Apparently, relatively cold and dry climate conditions persisted in NE Poland during the first ca. 1500 years of the Holocene, most likely due to a specific regional atmospheric circulation pattern. Prevailing anticyclonic circulation linked to a high-pressure cell above the remaining Scandinavian Ice Sheet (SIS) might have blocked the eastward propagation of warm and moist Westerlies and thus attenuated the early Holocene climatic amelioration in this region until the final decay of the SIS, a pattern different from climate development in Western Europe. The Lateglacial sediment record of Lake Mondsee (Upper Austria) was investigated in order to study the regional climate development and the environmental response to rapid climatic fluctuations. While the temperature rise and environmental response at the onset of the Holocene took place quasi-synchronously, major leads and lags in proxy responses characterize the onset of the Lateglacial Interstadial. In particular, the spread of coniferous woodlands and the reduction of detrital flux lagged the initial Lateglacial warming by ca. 500–750 years. Major cooling at the onset of the Younger Dryas took place synchronously with a change in vegetation, while the increase of detrital matter flux was delayed by about 150–300 years. Complex proxy responses are also detected for short-term Lateglacial climatic fluctuations. In summary, periods of abrupt climatic changes are characterized by complex and temporally variable proxy responses, mainly controlled by ecosystem inertia and the environmental preconditions. A second study on the Lake Mondsee sediment record focused on two small-scale climate deteriorations around 8200 and 9100 cal. a BP, which have been triggered by freshwater discharges to the North Atlantic, causing a shutdown of the Atlantic meridional overturning circulation (MOC). Combining microscopic varve counting and AMS 14C dating yielded a precise duration estimate (ca. 150 years) and absolute dating of the 8.2 ka cold event, both being in good agreement with results from other palaeoclimate records. Moreover, a sudden temperature overshoot after the 8.2 ka cold event was identified, also seen in other proxy records around the North Atlantic. This was most likely caused by enhanced resumption of the MOC, which also initiated substantial shifts of oceanic and atmospheric front systems. Although there is also evidence from other proxy records for pronounced recovery of the MOC and atmospheric circulation changes after the 9.1 ka cold event, no temperature overshoot is seen in the Lake Mondsee record, indicating the complex behaviour of the global climate system. The Holocene sediment record of Lake Iseo (northern Italy) was studied to shed light on regional earthquake activity and the influence of climate variability and anthropogenic impact on catchment erosion and detrital flux into the lake. Frequent small-scale detrital layers within the sediments reflect allochthonous sediment supply by extreme surface runoff events. During the early to mid-Holocene, increased detrital flux coincides with periods of cold and wet climate conditions, thus apparently being mainly controlled by climate variability. In contrast, intervals of high detrital flux during the late Holocene partly also correlate with phases of increased human impact, reflecting the complex influences on catchment erosion processes. Five large-scale event layers within the sediments, which are composed of mass-wasting deposits and turbidites, are supposed to have been triggered by strong local earthquakes. While the uppermost of these event layers is assigned to a documented adjacent earthquake in AD 1222, the four other layers are supposed to be related to previously undocumented prehistorical earthquakes.
The essay compares the dichotomous concepts of corporeality and spirituality in Judaism and Christianity. Through the ages, deviations from normative principles of beliefs could be discerned in both religions. These can be attributed either to the somewhat confrontational interaction between Jews and Christians in the Medieval urban environment or to the impact of Hellenic civilization on both monotheistic religions. Out of this dynamic impact emerged Christian art with a predilection to expressed corporeality, whereas Jewish religiosity found its artistic expression in a spiritual noniconographical mode. A genuine Jewish art and iconography could develop only after a certain degree of assimilation and secularization. Marc Chagall was the first protagonist of a mature expression of Jewish iconography.
In current practice, business processes modeling is done by trained method experts. Domain experts are interviewed to elicit their process information but not involved in modeling. We created a haptic toolkit for process modeling that can be used in process elicitation sessions with domain experts. We hypothesize that this leads to more effective process elicitation. This paper brakes down "effective elicitation" to 14 operationalized hypotheses. They are assessed in a controlled experiment using questionnaires, process model feedback tests and video analysis. The experiment compares our approach to structured interviews in a repeated measurement design. We executed the experiment with 17 student clerks from a trade school. They represent potential users of the tool. Six out of fourteen hypotheses showed significant difference due to the method applied. Subjects reported more fun and more insights into process modeling with tangible media. Video analysis showed significantly more reviews and corrections applied during process elicitation. Moreover, people take more time to talk and think about their processes. We conclude that tangible media creates a different working mode for people in process elicitation with fun, new insights and instant feedback on preliminary results.
What is visualization?
(2011)
Over the last 20 years, information visualization became a common tool in science and also a growing presence in the arts and culture at large. However, the use of visualization in cultural research is still in its infancy. Based on the work in the analysis of video games, cinema, TV, animation, Manga and other media carried out in Software Studies Initiative at University of California, San Diego over last two years, a number of visualization techniques and methods particularly useful for cultural and media research are presented.
Service-oriented Architectures (SOA) facilitate the provision and orchestration of business services to enable a faster adoption to changing business demands. Web Services provide a technical foundation to implement this paradigm on the basis of XML-messaging. However, the enhanced flexibility of message-based systems comes along with new threats and risks. To face these issues, a variety of security mechanisms and approaches is supported by the Web Service specifications. The usage of these security mechanisms and protocols is configured by stating security requirements in security policies. However, security policy languages for SOA are complex and difficult to create due to the expressiveness of these languages. To facilitate and simplify the creation of security policies, this thesis presents a model-driven approach that enables the generation of complex security policies on the basis of simple security intentions. SOA architects can specify these intentions in system design models and are not required to deal with complex technical security concepts. The approach introduced in this thesis enables the enhancement of any system design modelling languages – for example FMC or BPMN – with security modelling elements. The syntax, semantics, and notion of these elements is defined by our security modelling language SecureSOA. The metamodel of this language provides extension points to enable the integration into system design modelling languages. In particular, this thesis demonstrates the enhancement of FMC block diagrams with SecureSOA. To enable the model-driven generation of security policies, a domain-independent policy model is introduced in this thesis. This model provides an abstraction layer for security policies. Mappings are used to perform the transformation from our model to security policy languages. However, expert knowledge is required to generate instances of this model on the basis of simple security intentions. Appropriate security mechanisms, protocols and options must be chosen and combined to fulfil these security intentions. In this thesis, a formalised system of security patterns is used to represent this knowledge and to enable an automated transformation process. Moreover, a domain-specific language is introduced to state security patterns in an accessible way. On the basis of this language, a system of security configuration patterns is provided to transform security intentions related to data protection and identity management. The formal semantics of the security pattern language enable the verification of the transformation process introduced in this thesis and prove the correctness of the pattern application. Finally, our SOA Security LAB is presented that demonstrates the application of our model-driven approach to facilitate a dynamic creation, configuration, and execution of secure Web Service-based composed applications.
We show that the residue density of the logarithm of a generalized Laplacian on a closed manifold defines an invariant polynomial-valued differential form. We express it in terms of a finite sum of residues of
classical pseudodifferential symbols. In the case of the square of a Dirac operator, these formulas provide a pedestrian proof of the Atiyah–Singer formula for a pure Dirac operator in four dimensions and for a
twisted Dirac operator on a flat space of any dimension. These correspond to special cases of a more general formula by Scott and Zagier. In our approach, which is of perturbative nature, we use either a Campbell–Hausdorff formula derived by Okikiolu or a noncommutative Taylor-type formula.
Turning Aliens into Citizens
(2011)
Inhalt: Empirical results of the survey ; A cumulative index of citizenship ; Jammu and Kashmir: Contesting “Indian” citizenship ; Conclusion
Processes with independent increments are characterized via a duality formula, including Malliavin derivative and difference operators. This result is based on a characterization of infinitely divisible random vectors by a functional equation. A construction of the difference operator by a variational method is introduced and compared to approaches used by other authors for L´evy processes involving the chaos decomposition. Finally we extend our method to characterize infinitely divisible random measures.
It is well documented that transcriptionally coordinated genes tend to be functionally related, and that such relationships may be conserved across different species, and even kingdoms. (Ihmels et al., 2004). Such relationships was initially utilized to reveal functional gene modules in yeast and mammals (Ihmels et al., 2004), and to explore orthologous gene functions between different species and kingdoms (Stuart et al., 2003; Bergmann et al., 2004). Model organisms, such as Arabidopsis, are readily used in basic research due to resource availability and relative speed of data acquisition. A major goal is to transfer the acquired knowledge from these model organisms to species that are of greater importance to our society. However, due to large gene families in plants, the identification of functional equivalents of well characterized Arabidopsis genes in other plants is a non-trivial task, which often returns erroneous or inconclusive results. In this thesis, concepts of utilizing co-expression networks to help infer (i) gene function, (ii) organization of biological processes and (iii) knowledge transfer between species are introduced. An often overlooked fact by bioinformaticians is that a bioinformatic method is as useful as its accessibility. Therefore, majority of the work presented in this thesis was directed on developing freely available, user-friendly web-tools accessible for any biologist.
This co-authored paper is based on research that originated in 2003 when our team started a series of extensive field studies into the character of gameplay experiences. Originally within the Children as the Actors of Game Cultures research project, our aim was to better understand why particularly young people enjoy playing games, while also asking their parents how they perceive gaming as playing partners or as close observers. Gradually our in-depth interviews started to reveal a complex picture of more general relevance, where personal experiences, social contexts and cultural practices all came together to frame gameplay within something we called game cultures. Culture was the keyword, since we were not interested in studying games and play experiences in isolation, but rather as part of the rich meaning- making practices of lived reality.
We are interested in modeling some two-level population dynamics, resulting from the interplay of ecological interactions and phenotypic variation of individuals (or hosts) and the evolution of cells (or parasites) of two types living in these individuals. The ecological parameters of the individual dynamics depend on the number of cells of each type contained by the individual and the cell dynamics depends on the trait of the invaded individual. Our models are rooted in the microscopic description of a random (discrete) population of individuals characterized by one or several adaptive traits and cells characterized by their type. The population is modeled as a stochastic point process whose generator captures the probabilistic dynamics over continuous time of birth, mutation and death for individuals and birth and death for cells. The interaction between individuals (resp. between cells) is described by a competition between individual traits (resp. between cell types). We look for tractable large population approximations. By combining various scalings on population size, birth and death rates and mutation step, the single microscopic model is shown to lead to contrasting nonlinear macroscopic limits of different nature: deterministic approximations, in the form of ordinary, integro- or partial differential equations, or probabilistic ones, like stochastic partial differential equations or superprocesses. The study of the long time behavior of these processes seems very hard and we only develop some simple cases enlightening the difficulties involved.
The inspiral and merger of two black holes is among the most exciting and extreme events in our universe. Being one of the loudest sources of gravitational waves, they provide a unique dynamical probe of strong-field general relativity and a fertile ground for the observation of fundamental physics. While the detection of gravitational waves alone will allow us to observe our universe through an entirely new window, combining the information obtained from both gravitational wave and electro-magnetic observations will allow us to gain even greater insight in some of the most exciting astrophysical phenomena. In addition, binary black-hole mergers serve as an intriguing tool to study the geometry of space-time itself. In this dissertation we study the merger process of binary black-holes in a variety of conditions. Our results show that asymmetries in the curvature distribution on the common apparent horizon are correlated to the linear momentum acquired by the merger remnant. We propose useful tools for the analysis of black holes in the dynamical and isolated horizon frameworks and shed light on how the final merger of apparent horizons proceeds after a common horizon has already formed. We connect mathematical theorems with data obtained from numerical simulations and provide a first glimpse on the behavior of these surfaces in situations not accessible to analytical tools. We study electro-magnetic counterparts of super-massive binary black-hole mergers with fully 3D general relativistic simulations of binary black-holes immersed both in a uniform magnetic field in vacuum and in a tenuous plasma. We find that while a direct detection of merger signatures with current electro-magnetic telescopes is unlikely, secondary emission, either by altering the accretion rate of the circumbinary disk or by synchrotron radiation from accelerated charges, may be detectable. We propose a novel approach to measure the electro-magnetic radiation in these simulations and find a non-collimated emission that dominates over the collimated one appearing in the form of dual jets associated with each of the black holes. Finally, we provide an optimized gravitational wave detection pipeline using phenomenological waveforms for signals from compact binary coalescence and show that by including spin effects in the waveform templates, the detection efficiency is drastically improved as well as the bias on recovered source parameters reduced. On the whole, this disseration provides evidence that a multi-messenger approach to binary black-hole merger observations provides an exciting prospect to understand these sources and, ultimately, our universe.
Human-induced alterations of the environment are causing biotic changes worldwide, including the extinction of species and a mixing of once disparate floras and faunas. One type of biological communities that is expected to be particularly affected by environmental alterations are herb layer plant communities of fragmented forests such as those in the west European lowlands. However, our knowledge about current changes in species diversity and composition in these communities is limited due to a lack of adequate long-term studies. In this thesis, I resurveyed the herb layer communities of ancient forest patches in the Weser-Elbe region (NW Germany) after two decades using 175 semi-permanent plots. The general objectives were (i) to quantify changes in plant species diversity considering also between-community (β) and functional diversity, (ii) to determine shifts in species composition in terms of species’ niche breadth and functional traits and (iii) to find indications on the most likely environmental drivers for the observed changes. These objectives were pursued with four independent research papers (Chapters 1-4) whose results were brought together in a General Discussion. Alpha diversity (species richness) increased by almost four species on average, whereas β diversity tended to decrease (Chapter 1). The latter is interpreted as a beginning floristic homogenization. The observed changes were primarily the result of a spread of native habitat generalists that are able to tolerate broad pH and moisture ranges. The changes in α and β diversity were only significant when species abundances were neglected (Chapters 1 and 2), demonstrating that the diversity changes resulted mainly from gains and losses of low-abundance species. This study is one of the first studies in temperate Europe that demonstrates floristic homogenization of forest plant communities at a larger than local scale. The diversity changes found at the taxonomic level did not result in similar changes at the functional level (Chapter 2). The likely reason is that these communities are functionally “buffered”. Single communities involve most of the functional diversity of the regional pool, i.e., they are already functionally rich, while they are functionally redundant among each other, i.e., they are already homogeneous. Independent of taxonomic homogenization, the abundance of 30 species decreased significantly (Chapter 4). These species included 12 ancient forest species (i.e., species closely tied to forest patches with a habitat continuity > 200 years) and seven species listed on the Red List of endangered plant species in NW Germany. If these decreases continue over the next decades, local extinctions may result. This biotic impoverishment would seriously conflict with regional conservation goals. Community assembly mechanisms changed at the local level particularly at sites that experienced disturbance by forest management activities between the sampling periods (Chapter 3). Disturbance altered community assembly mechanisms in two ways: (i) it relaxed environmental filters and allowed the coexistence of different reproduction strategies, as reflected by a higher diversity of reproductive traits at the time of the resurvey, and (ii) it enhanced light availability and tightened competitive filters. These limited the functional diversity with respect to canopy height and selected for taller species. Thirty-one winner and 30 loser species, which had significantly increased or decreased in abundance, respectively, were characterized by various functional traits and ecological performances to find indications on the most likely environmental drivers for the observed floristic changes (Chapter 4). Winner species had higher seed longevity, flowered later in the season and had more often an oceanic distribution compared to loser species. Loser species tended to have a higher specific leaf area, to be more susceptible to deer browsing and to have a performance optimum at higher soil pH values compared to winner species. Multiple logistic regression analyses indicated that disturbances due to forest management interventions were the primary cause of the species shifts. As one of the first European resurvey studies, this study provides indications that an enhanced browsing pressure due to increased deer densities and increasingly warmer winters are important drivers. The study failed to demonstrate that eutrophication and acidification due to atmospheric deposition substantially drive herb layer changes. The restriction of the sample to the most base-rich sites in the region is discussed as a likely reason. Furthermore, the decline of several ancient forest species is discussed as an indication that the forest patches are still paying off their “extinction debt”, i.e., exhibit a delayed response to forest fragmentation.
IT systems for healthcare are a complex and exciting field. One the one hand, there is a vast number of improvements and work alleviations that computers can bring to everyday healthcare. Some ways of treatment, diagnoses and organisational tasks were even made possible by computer usage in the first place. On the other hand, there are many factors that encumber computer usage and make development of IT systems for healthcare a challenging, sometimes even frustrating task. These factors are not solely technology-related, but just as well social or economical conditions. This report describes some of the idiosyncrasies of IT systems in the healthcare domain, with a special focus on legal regulations, standards and security.
The lakes of the East African Rift System (EARS) have been intensively studied to better understand the influence of climate change on hydrological systems. The exceptional sensitivity of these rift lakes, however, is both a challenge and an opportunity when trying to reconstruct past climate changes from changes in the hydrological budget of lake basins on timescales 100 to 104 years. On one hand, differences in basin geometrics (shape, area, volume, depth), catchment rainfall distributions and varying erosion-deposition rates complicate regional interpretation of paleoclimate information from lacustrine sediment proxies. On the other hand, the sensitivity of rift lakes often provides paleoclimate records of excellent quality characterized by a high signal-to-noise ratio. This study aims at better understanding of the climate-proxy generating process in rift lakes by parameterizing the geomorphological and hydroclimatic conditions of a particular site providing a step towards the establishment of regional calibrations of transfer functions for climate reconstructions. The knowledge of the sensitivity of a lake basin to climate change furthermore is crucial for a better assessment of the probability of catastrophic changes in the future, which bear risks for landscapes, ecosystems, and organisms of all sorts, including humans. Part 1 of this thesis explores the effect of the morphology and the effective moisture of a lake catchment. The availability of digital elevation models (DEM) and gridded climate data sets facilitates the comparison of the morphological and hydroclimatic conditions of rift lakes. I used the hypsometric integral (HI) calculated from Shuttle Radar Topography Mission (SRTM) data to describe the morphology of ten lake basins in Kenya and Ethiopia. The aridity index (AI) describing the difference in the precipitation/evaporation balance within a catchment was used to compare the hydroclimatic of these basins. Correlating HI and AI with published Holocene lake-level variations revealed that lakes responding sensitively to relatively moderate climate change are typically graben shaped and characterized by a HI between 0.23-0.30, and relatively humid conditions with AI >1. These amplifier lakes, a term first introduced but not fully parameterized by Alayne Street-Perrott in the early 80s, are unexceptionally located in the crest of the Kenyan and Ethiopian domes. The non-amplifier lakes in the EARS either have lower HI 0.13-0.22 and higher AI (>1) or higher HI (0.31-0.37) and low AI (<1), reflecting pan-shaped morphologies with more arid hydroclimatic conditions. Part 2 of this work addresses the third important factor to be considered when using lake-level and proxy records to unravel past climate changes in the EARS: interbasin connectivity and groundwater flow through faulted and porous subsurface lithologies in a rift setting. First, I have compiled the available hydrogeological data including lithology, resistivity and water-well data for the adjacent Naivasha and Elmenteita-Nakuru basins in the Central Kenya Rift. Using this subsurface information and established records of lake-level decline at the last wet-dry climate transitions, i.e., the termination of the African Humid Period (AHP, 15 to 5 kyr BP), I used a linear decay model to estimate typical groundwater flow between the two basins. The results suggest a delayed response of the groundwater levels of ca. 5 kyrs if no recharge of groundwater occurs during the wet-dry transition, whereas the lag is 2-2.7 kyrs only using the modern recharge of ca. 0.52 m/yr. The estimated total groundwater flow from higher Lake Naivasha (1,880 m a.s.l. during the AHP) to Nakuru-Elmenteita (1,770 m) was 40 cubic kilometers. The unexpectedly large volume, more than half of the volume of the paleo-Lake Naivasha during the Early Holocene, emphasizes the importance of groundwater in hydrological modeling of paleo-lakes in rifts. Moreover, the subsurface connectivity of rift lakes also causes a significant lag time to the system introducing a nonlinear component to the system that has to be considered while interpreting paleo-lake records. Part 3 of this thesis investigated the modern intraseasonal precipitation variability within eleven lake basins discussed in the first section of the study excluding Lake Victoria and including Lake Tana. Remotely sensed rainfall estimates (RFE) from FEWS NET for 1996-2010, are used for the, March April May (MAM) July August September (JAS), October November (ON) and December January February (DJF). The seasonal precipitation are averaged and correlated with the prevailing regional and local climatic mechanisms. Results show high variability with Biennial to Triennial precipitation patterns. The spatial distribution of precipitation in JAS are linked to the onset and strength of the Congo Air Boundary (CAB) and Indian Summer Monsoon (ISM) dynamics. while in ON they are related to the strength of Positive ENSO and IOD phases This study describes the influence of graben morphologies, extreme climate constrasts within catchments and basins connectivity through faults and porous lithologies on rift lakes. Hence, it shows the importance of a careful characterization of a rift lake by these parameters prior to concluding from lake-level and proxy records to climate changes. Furthermore, this study highlights the exceptional sensitivity of rift lakes to relatively moderate climate change and its consequences for water availability to the biosphere including humans.
This volume contains the articles and papers which predominately have been published in international journals or edited volumes in the period from 1979 to 2009. The single articles reflect the main research areas of the editor and his co-authors who were engaged at the Kiel Institute of World Economics, the Johannes-Kepler-University Linz/Austria, the Justus- Liebig-University Giessen, the University of Potsdam, and the German Institute for Economic Research (DIW Berlin).
This volume contains the articles and papers which predominately have been published in international journals or edited volumes in the period from 1979 to 2009. The single articles reflect the main research areas of the editor and his co-authors who were engaged at the Kiel Institute of World Economics, the Johannes-Kepler-University Linz/Austria, the Justus- Liebig-University Giessen, the University of Potsdam, and the German Institute for Economic Research (DIW Berlin).
This volume contains the articles and papers which predominately have been published in international journals or edited volumes in the period from 1979 to 2009. The single articles reflect the main research areas of the editor and his co-authors who were engaged at the Kiel Institute of World Economics, the Johannes-Kepler-University Linz/Austria, the Justus- Liebig-University Giessen, the University of Potsdam, and the German Institute for Economic Research (DIW Berlin).
This volume contains the articles and papers which predominately have been published in international journals or edited volumes in the period from 1979 to 2009. The single articles reflect the main research areas of the editor and his co-authors who were engaged at the Kiel Institute of World Economics, the Johannes-Kepler-University Linz/Austria, the Justus- Liebig-University Giessen, the University of Potsdam, and the German Institute for Economic Research (DIW Berlin).
This volume contains the articles and papers which predominately have been published in international journals or edited volumes in the period from 1979 to 2009. The single articles reflect the main research areas of the editor and his co-authors who were engaged at the Kiel Institute of World Economics, the Johannes-Kepler-University Linz/Austria, the Justus- Liebig-University Giessen, the University of Potsdam, and the German Institute for Economic Research (DIW Berlin).
This volume contains the articles and papers which predominately have been published in international journals or edited volumes in the period from 1979 to 2009. The single articles reflect the main research areas of the editor and his co-authors who were engaged at the Kiel Institute of World Economics, the Johannes-Kepler-University Linz/Austria, the Justus- Liebig-University Giessen, the University of Potsdam, and the German Institute for Economic Research (DIW Berlin).
Conventional energy sources are diminishing and non-renewable, take million years to form and cause environmental degradation. In the 21st century, we have to aim at achieving sustainable, environmentally friendly and cheap energy supply by employing renewable energy technologies associated with portable energy storage devices. Lithium-ion batteries can repeatedly generate clean energy from stored materials and convert reversely electric into chemical energy. The performance of lithium-ion batteries depends intimately on the properties of their materials. Presently used battery electrodes are expensive to be produced; they offer limited energy storage possibility and are unsafe to be used in larger dimensions restraining the diversity of application, especially in hybrid electric vehicles (HEVs) and electric vehicles (EVs). This thesis presents a major progress in the development of LiFePO4 as a cathode material for lithium-ion batteries. Using simple procedure, a completely novel morphology has been synthesized (mesocrystals of LiFePO4) and excellent electrochemical behavior was recorded (nanostructured LiFePO4). The newly developed reactions for synthesis of LiFePO4 are single-step processes and are taking place in an autoclave at significantly lower temperature (200 deg. C) compared to the conventional solid-state method (multi-step and up to 800 deg. C). The use of inexpensive environmentally benign precursors offers a green manufacturing approach for a large scale production. These newly developed experimental procedures can also be extended to other phospho-olivine materials, such as LiCoPO4 and LiMnPO4. The material with the best electrochemical behavior (nanostructured LiFePO4 with carbon coating) was able to delive a stable 94% of the theoretically known capacity.
The complete consumption of the oceanic domain of a tectonic plate by subduction into the upper mantle results in continent subduction, although continental crust is typically of lower density than the upper mantle. Thus, the sites of former oceanic domains (named suture zones) are generally decorated with stratigraphic sequences deposited along continental passive margins that were metamorphosed under low-grade, high-pressure conditions, i.e., low temperature/depth ratios (< 15°C/km) with respect to geothermal gradients in tectonically stable regions. Throughout the Mesozoic and Cenozoic (i.e., since ca. 250 Ma), the Mediterranean realm was shaped by the closure of the Tethyan Ocean, which likely consisted in numerous oceanic domains and microcontinents. However, the exact number and position of Tethyan oceans and continents (i.e., the Tethyan palaeogeography) remains debated. This is particularly the case of Western and Central Anatolia, where a continental fragment was accreted to the southern composite margin of the Eurasia sometime between the Late Cretaceous and the early Cenozoic. The most frontal part of this microcontinent experienced subduction-related metamorphism around 85-80 Ma, and collision-related metamorphism affected more external parts around 35 Ma. This unsually-long period between subduction- and collision-related metamorphisms (ca. 50 Ma) in units ascribed to the same continental edge constitutes a crucial issue to address in order to unravel how Anatolia was assembled. The Afyon Zone is a tectono-sedimentary unit exposed south and structurally below the front high-pressure belt. It is composed of a Mesozoic sedimentary sequence deposited on top of a Precambrian to Palaeozoic continental substratum, which can be traced from Northwestern to southern Central Anatolia, along a possible Tethyan suture. Whereas the Afyon Zone was defined as a low-pressure metamorphic unit, high-pressure minerals (mainly Fe-Mg-carpholite in metasediments) were recently reported from its central part. These findings shattered previous conceptions on the tectono-metamorphic evolution of the Afyon Zone in particular, and of the entire region in general, and shed light on the necessity to revise the regional extent of subduction-related metamorphism by re-inspecting the petrology of poorly-studied metasediments. In this purpose, I re-evaluated the metamorphic evolution of the entire Afyon Zone starting from field observations. Low-grade, high-pressure mineral assemblages (Fe-Mg-carpholite and glaucophane) are reported throughout the unit. Well-preserved carpholite-chloritoid assemblages are useful to improve our understanding of mineral relations and transitions in the FeO-MgO-Al2O3-SiO2-H2O system during rocks’ travel down to depth (prograde metamorphism). Inspection of petrographic textures, minute variations in mineral composition and Mg-Fe distribution among carpholite-chloritoid assemblages documents multistage mineral growth, accompanied by a progressive enrichment in Mg, and strong element partitioning. Using an updated database of mineral thermodynamic properties, I modelled the pressure and temperature conditions that are consistent with textural and chemical observations. Carpholite-bearing assemblages in the Afyon Zone account for a temperature increase from 280 to 380°C between 0.9 and 1.1 GPa (equivalent to a depth of 30-35 km). In order to further constrain regional geodynamics, first radiometric ages were determined in close association with pressure-temperature estimates for the Afyon Zone, as well as two other tectono-sedimentary units from the same continental passive margin (the Ören and Kurudere-Nebiler Units from SW Anatolia). For age determination, I employed 40Ar-39Ar geochronology on white mica in carpholite-bearing rocks. For thermobarometry, a multi-equilibrium approach was used based on quartz-chlorite-mica and quartz-chlorite-chloritoid associations formed at the expense of carpholite-bearing assemblages, i.e., during the exhumation from the subduction zone. This combination allows deciphering the significance of the calculated radiometric ages in terms of metamorphic conditions. Results show that the Afyon Zone and the Ören Unit represent a latest Cretaceous high-pressure metamorphic belt, and the Kurudere-Nebiler Unit was affected by subduction-related metamorphism around 45 Ma and cooled down after collision-related metamorphism around 26 Ma. The results provided in the present thesis and from the literature allow better understanding continental amalgamation in Western Anatolia. It is shown that at least two distinct oceanic branches, whereas only one was previously considered, have closed during continuous north-dipping subduction between 92 and 45 Ma. Between 85-80 and 70-65 Ma, a narrow continental domain (including the Afyon Zone) was buried into a subduction zone within the northern oceanic strand. Parts of the subducted continent crust were exhumed while the upper oceanic plate was transported southwards. Subduction of underlying lithosphere persisted, leading to the closure of the southern oceanic branch and to subduct the front of a second continental domain (including the Kurudere-Nebiler Unit). This followed by a continental collisional stage characterized by the cease of subduction, crustal thicknening and the detachment of the subducting oceanic slab from the accreted continent lithosphere. The present study supports that in the late Mesozoic the East Mediterranean realm had a complex tectonic configuration similar to present Southeast Asia or the Caribbean, with multiple, coexisting oceanic basins, microcontinents and subduction zones.