Refine
Year of publication
- 2014 (222) (remove)
Document Type
- Doctoral Thesis (222) (remove)
Keywords
- Gammastrahlungsastronomie (4)
- data analysis (4)
- gamma-ray astronomy (4)
- Crab Nebula (3)
- Datenanalyse (3)
- Krebsnebel (3)
- Synchronisation (3)
- Systembiologie (3)
- synchronization (3)
- systems biology (3)
Institute
- Institut für Biochemie und Biologie (37)
- Institut für Physik und Astronomie (25)
- Institut für Geowissenschaften (19)
- Wirtschaftswissenschaften (15)
- Institut für Chemie (13)
- Historisches Institut (12)
- Institut für Informatik und Computational Science (10)
- Department Psychologie (9)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (9)
- Institut für Ernährungswissenschaft (8)
Indonesien zählt zu den weltweit führenden Ländern bei der Nutzung von geothermischer Energie. Die geothermischen Energiequellen sind im Wesentlichen an den aktiven Vulkanismus gebunden, der durch die Prozesse an der indonesischen Subduktionszone verursacht wird. Darüber hinaus sind geotektonische Strukturen wie beispielsweise die Sumatra-Störung als verstärkende Faktoren für das geothermische Potenzial von Bedeutung. Bei der geophysikalischen Erkundung der indonesischen Geothermie-Ressourcen konzentrierte man sich bisher vor allem auf die Magnetotellurik. Passive Seismologie wurde dahingegen ausschließlich für die Überwachung von im Betrieb befindlichen Geothermie-Anlagen verwendet. Jüngste Untersuchungungen z.B. in Island und in den USA haben jedoch gezeigt, dass seismologische Verfahren bereits in der Erkundungsphase wichtige Informationen zu den physikalischen Eigenschaften, zum Spannungsfeld und zu möglichen Fluid- und Wärmetransportwegen liefern können. In der vorgelegten Doktorarbeit werden verschiedene moderne Methoden der passiven Seismologie verwendet, um beispielhaft ein neues, von der indonesischen Regierung für zukünftige geothermische Energiegewinnung ausgewiesenes Gebiet im nördlichen Teil Sumatras (Indonesien) zu erkunden. Die konkreten Ziele der Untersuchungen umfassten (1) die Ableitung von 3D Strukturmodellen der P- und S-Wellen Geschwindigkeiten (Parameter Vp und Vs), (2) die Bestimmung der Absorptionseigenschaften (Parameter Qp), und (3) die Kartierung und Charakterisierung von Störungssystemen auf der Grundlage der Seismizitätsverteilung und der Herdflächenlösungen. Für diese Zwecke habe ich zusammen mit Kollegen ein seismologisches Netzwerk in Tarutung (Sumatra) aufgebaut und über einen Zeitraum von 10 Monaten (Mai 2011 – Februar 2012) betrieben. Insgesamt wurden hierbei 42 Stationen (jeweils ausgestattet mit EDL-Datenlogger, 3-Komponenten, 1 Hz Seismometer) über eine Fläche von etwa 35 x 35 km verteilt. Mit dem Netzwerk wurden im gesamten Zeitraum 2568 lokale Erdbeben registriert. Die integrierte Betrachtung der Ergebnisse aus den verschiedenen Teilstudien (Tomographie, Erdbebenverteilung) erlaubt neue Einblicke in die generelle geologische Stukturierung sowie eine Eingrenzung von Bereichen mit einem erhöhten geothermischen Potenzial. Das tomographische Vp-Modell ermöglicht eine Bestimmung der Geometrie von Sedimentbecken entlang der Sumatra-Störung. Für die Geothermie besonders interessant ist der Bereich nordwestlich des Tarutung-Beckens. Die dort abgebildeten Anomalien (erhöhtes Vp/Vs, geringes Qp) habe ich als mögliche Aufstiegswege von warmen Fluiden interpretiert. Die scheinbar asymetrische Verteilung der Anomalien wird hierbei im Zusammenhang mit der Seismizitätsverteilung, der Geometrie der Beben-Bruchflächen, sowie struktur-geologischen Modellvorstellungen diskutiert. Damit werden wesentliche Informationen für die Planung einer zukünftigen geothermischen Anlage bereitgestellt.
Bacteria respond to changing environmental conditions by switching the global pattern of expressed genes. In response to specific environmental stresses the cell activates several stress-specific molecules such as sigma factors. They reversibly bind the RNA polymerase to form the so-called holoenzyme and direct it towards the appropriate stress response genes. In exponentially growing E. coli cells, the majority of the transcriptional activity is carried out by the housekeeping sigma factor, while stress responses are often under the control of alternative sigma factors. Different sigma factors compete for binding to a limited pool of RNA polymerase (RNAP) core enzymes, providing a mechanism for cross talk between genes or gene classes via the sharing of expression machinery. To quantitatively analyze the contribution of sigma factor competition to global changes in gene expression, we develop a thermodynamic model that describes binding between sigma factors and core RNAP at equilibrium, transcription, non-specific binding to DNA and the modulation of the availability of the molecular components.
Association of housekeeping sigma factor to RNAP is generally favored by its abundance and higher binding affinity to the core. In order to promote transcription by alternative sigma subunits, the bacterial cell modulates the transcriptional efficiency in a reversible manner through several strategies such as anti-sigma factors, 6S RNA and generally any kind of transcriptional regulators (e.g. activators or inhibitors). By shifting the outcome of sigma factor competition for the core, these modulators bias the transcriptional program of the cell. The model is validated by comparison with in vitro competition experiments, with which excellent agreement is found. We observe that transcription is affected via the modulation of the concentrations of the different types of holoenzymes, so saturated promoters are only weakly affected by sigma factor competition. However, in case of overlapping promoters or promoters recognized by two types of sigma factors, we find that even saturated promoters are strongly affected.
Active transcription effectively lowers the affinity between the sigma factor driving it and the core RNAP, resulting in complex cross talk effects and raising the question of how their in vitro measure is relevant in the cell. We also estimate that sigma factor competition is not strongly affected by non-specific binding of core RNAPs, sigma factors, and holoenzymes to DNA. Finally, we analyze the role of increased core RNAP availability upon the shut-down of ribosomal RNA transcription during stringent response. We find that passive up-regulation of alternative sigma-dependent transcription is not only possible, but also displays hypersensitivity based on the sigma factor competition. Our theoretical analysis thus provides support for a significant role of passive control during that global switch of the gene expression program and gives new insights into RNAP partitioning in the cell.
Con la sua proposta di una Somaestetica, articolata fondamentalmente in analitica, pragmatica e pratica, Richard Shusterman intende in primo luogo fornire e creare una cornice metodologica, un orientamento unitario che sia in grado di rintracciare, ricostruire e portare a manifestazione - all’interno di eterogenee riflessioni teoriche e pratiche somatiche - la comune esigenza di ridare luce alla dimensione corporea come modo primario di essere nel mondo. Recuperando l’accezione baumgarteniana di Aesthetica come gnoseologia inferiore, arte dell’analogo della ragione, scienza della conoscenza sensibile, la somaestetica intende dare nuovo impulso alla più profonda radice di estetica e filosofia che coglie la vita nel suo processo di metamorfosi e rigenerazione continua, in quel respiro vitale che, per quanto possa diventare cosciente, non è mai totalmente afferrabile dalla ragione discorsiva, situandosi piuttosto in quello spazio primordiale in cui coscienza e corpo si coappartengono, in cui il soggetto non è ancora individualizzabile perché fuso con l’ambiente, non è totalmente privatizzabile perché intrinsecamente plasmato dal tessuto sociale cui egli stesso conferisce dinamicamente forma. A partire dunque dalla rivalutazione del concetto di Aisthesis la disciplina somaestetica mira ad una intensificazione di sensorialità, percezione, emozione, commozione, rintracciando proprio nel Soma la fonte di quelle facoltà “inferiori” irriducibili a quelle puramente intellettuali, che permettono di accedere alle dimensioni qualitative dell’esperienza, di portare a manifestazione e far maturare l’essere umano come essere indivisibile che non si lascia incontrare da un pensiero che ne rinnega l’unitarietà in nome di fittizie e laceranti distinzioni dicotomiche. Nel corpo infatti si radicano in modo silente regole, convenzioni, norme e valori socioculturali che determinano e talvolta limitano la configurazione ed espressione di sensazioni, percezioni, cognizioni, pensieri, azioni, volizioni, disposizioni di un soggetto da sempre inserito in una Mitwelt (mondo comune), ed è allora proprio al corpo che bisogna rivolgersi per riconfigurare più autentiche modalità di espressione del soggetto che crea equilibri dinamici per mantenere una relazione di coerenza con il più ampio contesto sociale, culturale, ambientale. L’apertura al confronto con eterogenee posizioni filosofiche e l’intrinseca multidisciplinarietà spiegano la centralità nel contemporaneo dibattito estetologico internazionale della Somaestetica che, rivolgendosi tanto ad una formulazione teorica quanto ad una concreta applicazione pratica, intende rivalutare il soma come corporeità intelligente, senziente, intenzionale e attiva, non riducibile all’accezione peccaminosa di caro (mero corpo fisico privo di vita e sensazione). Attraverso la riflessione e la pratica di tecniche di coscienza somatica si portano in primo piano i modi in cui il sempre più consapevole rapporto con la propria corporeità come mediatamente esperita e immediatamente vissuta, sentita, offre occasioni autentiche di realizzazione progressiva di sé innanzitutto come persone, capaci di autocoltivazione, di riflessione cosciente sulle proprie abitudini incorporate, di ristrutturazione creativa di sé, di intensificata percezione e apprezzamento sensoriale sia nel concreto agire quotidiano, sia nella dimensione più propriamente estetologica di ricezione, fruizione e creazione artistica. L’indirizzo essenzialmente pragmatista della riflessione di Shusterman traccia così una concezione fondamentalmente relazionale dell’estetica in grado di porsi proprio nel movimento e nel rapporto continuamente diveniente di vera e propria trasformazione e passaggio tra le dimensioni fisiche, proprio-corporee, psichiche e spirituali del soggetto la cui interazione, ed il cui reciproco riversarsi le une nelle altre, può risultare profondamente arricchito attraverso una progressiva e sempre crescente consapevolizzazione della ricchezza della dimensione corporea in quanto intenzionale, percettiva, senziente, volitiva, tanto quanto vulnerabile, limitante, caduca, patica. Il presente lavoro intende ripercorrere ed approfondire alcuni dei principali referenti di Shusterman, focalizzandosi prevalentemente sulla radice pragmatista della sua proposta e sul confronto con il dibattito di area tedesca tra estetica, antropologia filosofica, neofenomenologia e antropologia medica, per riguadagnare una nozione di soma che proprio a partire dal contrasto, dall’impatto irriducibile con la potenza annullante delle situazioni limite, della crisi possa acquisire un più complesso e ricco valore armonizzante delle intrinseche e molteplici dimensioni che costituiscono il tessuto della soggettività incarnata. In particolare il primo capitolo (1. Somaestetica) chiarisce le radici essenzialmente pragmatiste della proposta shustermaniana e mostra come sia possibile destrutturare e dunque riconfigurare radicati modi di esperienza, rendendo coscienti abitudini e modi di vivere che si fissano a livello somatico in modo per lo più inavvertito. Il confronto con la nozione di Habitus, di cui Pierre Bourdieu mette brillantemente in luce l’invisibile e socialmente determinata matrice somatica, lascia scorgere come ogni manifestazione umana sia sostenuta dall’incorporazione di norme, credenze, valori che determinano e talvolta limitano l’espressione, lo sviluppo, persino le predisposizioni e le inclinazioni degli individui. Ed è proprio intervenendo a questo livello che si può restituire libertà alle scelte e aprirsi così alle dimensioni essenzialmente qualitative dell’esperienza che, nell’accezione deweyana è un insieme olistico unitario e coeso che fa da sfondo alle relazioni organismo-ambiente, un intreccio inestricabile di teoria e prassi, particolare e universale, psiche e soma, ragione ed emozione, percettivo e concettuale, insomma quell’immediata conoscenza corporea che struttura lo sfondo di manifestazione della coscienza.
Today, it is well known that galaxies like the Milky Way consist not only of stars but also of gas and dust. The galactic halo, a sphere of gas that surrounds the stellar disk of a galaxy, is especially interesting. It provides a wealth of information about in and outflowing gaseous material towards and away from galaxies and their hierarchical evolution. For the Milky Way, the so-called high-velocity clouds (HVCs), fast moving neutral gas complexes in the halo that can be traced by absorption-line measurements, are believed to play a crucial role in the overall matter cycle in our Galaxy. Over the last decades, the properties of these halo structures and their connection to the local circumgalactic and intergalactic medium (CGM and IGM, respectively) have been investigated in great detail by many different groups. So far it remains unclear, however, to what extent the results of these studies can be transferred to other galaxies in the local Universe. In this thesis, we study the absorption properties of Galactic HVCs and compare the HVC absorption characteristics with those of intervening QSO absorption-line systems at low redshift. The goal of this project is to improve our understanding of the spatial extent and physical conditions of gaseous galaxy halos in the local Universe. In the first part of the thesis we use HST /STIS ultraviolet spectra of more than 40 extragalactic background sources to statistically analyze the absorption properties of the HVCs in the Galactic halo. We determine fundamental absorption line parameters including covering fractions of different weakly/intermediately/highly ionized metals with a particular focus on SiII and MgII. Due to the similarity in the ionization properties of SiII and MgII, we are able to estimate the contribution of HVC-like halo structures to the cross section of intervening strong MgII absorbers at z = 0. Our study implies that only the most massive HVCs would be regarded as strong MgII absorbers, if the Milky Way halo would be seen as a QSO absorption line system from an exterior vantage point. Combining the observed absorption-cross section of Galactic HVCs with the well-known number density of intervening strong MgII absorbers at z = 0, we conclude that the contribution of infalling gas clouds (i.e., HVC analogs) in the halos of Milky Way-type galaxies to the cross section of strong MgII absorbers is 34%. This result indicates that only about one third of the strong MgII absorption can be associated with HVC analogs around other galaxies, while the majority of the strong MgII systems possibly is related to galaxy outflows and winds. The second part of this thesis focuses on the properties of intervening metal absorbers at low redshift. The analysis of the frequency and physical conditions of intervening metal systems in QSO spectra and their relation to nearby galaxies offers new insights into the typical conditions of gaseous galaxy halos. One major aspect in our study was to regard intervening metal systems as possible HVC analogs. We perform a detailed analysis of absorption line properties and line statistics for 57 metal absorbers along 78 QSO sightlines using newly-obtained ultraviolet spectra obtained with HST /COS. We find clear evidence for bimodal distribution in the HI column density in the absorbers, a trend that we interpret as sign for two different classes of absorption systems (with HVC analogs at the high-column density end). With the help of the strong transitions of SiII λ1260, SiIII λ1206, and CIII λ977 we have set up Cloudy photoionization models to estimate the local ionization conditions, gas densities, and metallicities. We find that the intervening absorption systems studied by us have, on average, similar physical conditions as Galactic HVC absorbers, providing evidence that many of them represent HVC analogs in the vicinity of other galaxies. We therefore determine typical halo sizes for SiII, SiIII, and CIII for L = 0.01L∗ and L = 0.05L∗ galaxies. Based on the covering fractions of the different ions in the Galactic halo, we find that, for example, the typical halo size for SiIII is ∼ 160 kpc for L = 0.05L∗ galaxies. We test the plausibility of this result by searching for known galaxies close to the QSO sightlines and at similar redshifts as the absorbers. We find that more than 34% of the measured SiIII absorbers have galaxies associated with them, with the majority of the absorbers indeed being at impact parameters ρ ≤160 kpc.
When azobenzene-modified photosensitive polymer films are irradiated with light interference patterns, topographic variations in the film develop that follow the electric field vector distribution resulting in the formation of surface relief grating (SRG). The exact correspondence of the electric field vector orientation in interference pattern in relation to the presence of local topographic minima or maxima of SRG is in general difficult to determine. In my thesis, we have established a systematic procedure to accomplish the correlation between different interference patterns and the topography of SRG. For this, we devise a new setup combining an atomic force microscope and a two-beam interferometer (IIAFM). With this set-up, it is possible to track the topography change in-situ, while at the same time changing polarization and phase of the impinging interference pattern. To validate our results, we have compared two photosensitive materials named in short as PAZO and trimer. This is the first time that an absolute correspondence between the local distribution of electric field vectors of interference pattern and the local topography of the relief grating could be established exhaustively. In addition, using our IIAFM we found that for a certain polarization combination of two orthogonally polarized interfering beams namely SP (↕, ↔) interference pattern, the topography forms SRG with only half the period of the interference patterns. Exploiting this phenomenon we are able to fabricate surface relief structures below diffraction limit with characteristic features measuring only 140 nm, by using far field optics with a wavelength of 491 nm. We have also probed for the stresses induced during the polymer mass transport by placing an ultra-thin gold film on top (5–30 nm). During irradiation, the metal film not only deforms along with the SRG formation, but ruptures in regular and complex manner. The morphology of the cracks differs strongly depending on the electric field distribution in the interference pattern even when the magnitude and the kinetic of the strain are kept constant. This implies a complex local distribution of the opto-mechanical stress along the topography grating. The neutron reflectivity measurements of the metal/polymer interface indicate the penetration of metal layer within the polymer resulting in the formation of bonding layer that confirms the transduction of light induced stresses in the polymer layer to a metal film.
The atmosphere over the Arctic Ocean is strongly influenced by the distribution of sea ice and open water. Leads in the sea ice produce strong convective fluxes of sensible and latent heat and release aerosol particles into the atmosphere. They increase the occurrence of clouds and modify the structure and characteristics of the atmospheric boundary layer (ABL) and thereby influence the Arctic climate.
In the course of this study aircraft measurements were performed over the western Arctic Ocean as part of the campaign PAMARCMIP 2012 of the Alfred Wegener Institute for Polar and Marine Research (AWI). Backscatter from aerosols and clouds within the lower troposphere and the ABL were measured with the nadir pointing Airborne Mobile Aerosol Lidar (AMALi) and dropsondes were launched to obtain profiles of meteorological variables. Furthermore, in situ measurements of aerosol properties, meteorological variables and turbulence were part of the campaign. The measurements covered a broad range of atmospheric and sea ice conditions.
In this thesis, properties of the ABL over Arctic sea ice with a focus on the influence of open leads are studied based on the data from the PAMARCMIP campaign. The height of the ABL is determined by different methods that are applied to dropsonde and AMALi backscatter profiles. ABL heights are compared for different flights representing different conditions of the atmosphere and of sea ice and open water influence. The different criteria for ABL height that are applied show large variation in terms of agreement among each other, depending on the characteristics of the ABL and its history. It is shown that ABL height determination from lidar backscatter by methods commonly used under mid-latitude conditions is applicable to the Arctic ABL only under certain conditions. Aerosol or clouds within the ABL are needed as a tracer for ABL height detection from backscatter. Hence an aerosol source close to the surface is necessary, that is typically found under the present influence of open water and therefore convective conditions. However it is not always possible to distinguish residual layers from the actual ABL. Stable boundary layers are generally difficult to detect.
To illustrate the complexity of the Arctic ABL and processes therein, four case studies are analyzed each of which represents a snapshot of the interplay between atmosphere and underlying sea ice or water surface. Influences of leads and open water on the aerosol and clouds within the ABL are identified and discussed. Leads are observed to cause the formation of fog and cloud layers within the ABL by humidity emission. Furthermore they decrease the stability and increase the height of the ABL and consequently facilitate entrainment of air and aerosol layers from the free troposphere.
NutzerInnen von gewalthaltigen Medien geben einerseits oftmals zu, dass sie fiktionale, gewalthaltige Medien konsumieren, behaupten jedoch gleichzeitig, dass dies nicht ihr Verhalten außerhalb des Medienkontexts beeinflusst. Sie argumentieren, dass sie leicht zwischen Dingen, die im fiktionalen Kontext und Dingen, die in der Realität gelernt wurden, unterscheiden können. Im Kontrast zu diesen Aussagen zeigen Metanalysen Effektstärken im mittleren Bereich für den Zusammenhang zwischen Gewaltmedienkonsum und aggressivem Verhalten. Diese Ergebnisse können nur erklärt werden, wenn MediennutzerInnen gewalthaltige Lernerfahrungen auch außerhalb des Medienkontexts anwenden. Ein Prozess, der Lernerfahrungen innerhalb des Medienkontexts mit dem Verhalten in der realen Welt verknüpft, ist Desensibilisierung, die oftmals eine Reduktion des negativen Affektes gegenüber Gewalt definiert ist. Zur Untersuchung des Desensibilisierungsprozesses wurden vier Experimente durchgeführt. Die erste in dieser Arbeit untersuchte Hypothese war, dass je häufiger Personen Gewaltmedien konsumieren, desto weniger negativen Affekt zeigen sie gegenüber Bildern mit realer Gewalt. Jedoch wurde angenommen, dass diese Bewertung auf Darstellungen von realer Gewalt beschränkt ist und nicht bei Bildern ohne Gewaltbezug, die einen negativen Affekt auslösen, zu finden ist. Die zweite Hypothese bezog sich auf den Affekt während des Konsums von Mediengewalt. Hier wurde angenommen, dass besonders Personen, die Freude an Gewalt in den Medien empfinden weniger negativen Affekt gegenüber realen Gewaltdarstellungen zeigen. Die letzte Hypothese beschäftigte sich mit kognitiver Desensibilisierung und sagte vorher, dass Gewaltmedienkonsum zu einem Transfer von Reaktionen, die normalerweise gegenüber gewalthaltigen Reizen gezeigt werden, auf ursprünglich neutrale Reize führt. Das erste Experiment (N = 57) untersuchte, ob die habituelle Nutzung von gewalthaltigen Medien den selbstberichteten Affekt (Valenz und Aktivierung) gegenüber Darstellungen von realer Gewalt und nichtgewalthaltigen Darstellungen, die negativen Affekt auslösen, vorhersagt. Die habituelle Nutzung von gewalthaltigen Medien sagte weniger negative Valenz und weniger allgemeine Aktivierung gegenüber gewalthalten und nichtgewalthaltigen Bildern vorher. Das zweite Experiment (N = 103) untersuchte auch die Beziehung zwischen habituellem Gewaltmedienkonsum und den affektiven Reaktionen gegenüber Bildern realer Gewalt und negativen affektauslösenden Bildern. Als weiterer Prädiktor wurde der Affekt beim Betrachten von gewalthaltigen Medien hinzugefügt. Der Affekt gegenüber den Bildern wurde zusätzlich durch psychophysiologische Maße (Valenz: C: Supercilii; Aktivierung: Hautleitreaktion) erhoben. Wie zuvor sagte habitueller Gewaltmedienkonsum weniger selbstberichte Erregung und weniger negative Valenz für die gewalthaltigen und die negativen, gewalthaltfreien Bilder vorher. Die physiologischen Maßen replizierten dieses Ergebnis. Jedoch zeigte sich ein anderes Muster für den Affekt beim Konsum von Gewalt in den Medien. Personen, die Gewalt in den Medien stärker erfreut, zeigen eine Reduktion der Responsivität gegenüber Gewalt auf allen vier Maßen. Weiterhin war bei drei dieser vier Maße (selbstberichte Valenz, Aktivität des C. Supercilii und Hautleitreaktion) dieser Zusammenhang auf die gewalthaltigen Bilder beschränkt, mit keinem oder nur einem kleinen Effekt auf die negativen, aber nichtgewalthaltigen Bilder. Das dritte Experiment (N = 73) untersuchte den Affekt während die Teilnehmer ein Computerspiel spielten. Das Spiel wurde eigens für dieses Experiment programmiert, sodass einzelne Handlungen im Spiel mit der Aktivität des C. Supercilii, dem Indikator für negativen Affekt, in Bezug gesetzt werden konnten. Die Analyse des C. Supercilii zeigte, dass wiederholtes Durchführen von aggressiven Spielzügen zu einem Rückgang von negativen Affekt führte, der die aggressiven Spielhandlungen begleitete. Der negative Affekt während gewalthaltiger Spielzüge wiederum sagte die affektive Reaktion gegenüber Darstellungen von gewalthaltigen Bildern vorher, nicht jedoch gegenüber den negativen Bildern. Das vierte Experiment (N = 77) untersuchte kognitive Desensibilisierung, die die Entwicklung von Verknüpfungen zwischen neutralen und aggressiven Kognitionen beinhaltete. Die Teilnehmer spielten einen Ego-Shooter entweder auf einem Schiff- oder einem Stadtlevel. Die Beziehung zwischen den neutralen Konstrukten (Schiff/Stadt) und den aggressiven Kognitionen wurde mit einer lexikalischen Entscheidungsaufgabe gemessen. Das Spielen im Schiff-/Stadt-Level führte zu einer kürzen Reaktionszeit für aggressive Wörter, wenn sie einem Schiff- bzw. Stadtprime folgten. Dies zeigte, dass die im Spiel enthaltenen neutralen Konzepte mit aggressiven Knoten verknüpft werden. Die Ergebnisse dieser vier Experimente wurden diskutiert im Rahmen eines lerntheoretischen Ansatzes um Desensibilisierung zu konzeptualisieren.
Boolean constraint solving technology has made tremendous progress over the last decade, leading to industrial-strength solvers, for example, in the areas of answer set programming (ASP), the constraint satisfaction problem (CSP), propositional satisfiability (SAT) and satisfiability of quantified Boolean formulas (QBF). However, in all these areas, there exist multiple solving strategies that work well on different applications; no strategy dominates all other strategies. Therefore, no individual solver shows robust state-of-the-art performance in all kinds of applications. Additionally, the question arises how to choose a well-performing solving strategy for a given application; this is a challenging question even for solver and domain experts. One way to address this issue is the use of portfolio solvers, that is, a set of different solvers or solver configurations. We present three new automatic portfolio methods: (i) automatic construction of parallel portfolio solvers (ACPP) via algorithm configuration,(ii) solving the $NP$-hard problem of finding effective algorithm schedules with Answer Set Programming (aspeed), and (iii) a flexible algorithm selection framework (claspfolio2) allowing for fair comparison of different selection approaches. All three methods show improved performance and robustness in comparison to individual solvers on heterogeneous instance sets from many different applications. Since parallel solvers are important to effectively solve hard problems on parallel computation systems (e.g., multi-core processors), we extend all three approaches to be effectively applicable in parallel settings. We conducted extensive experimental studies different instance sets from ASP, CSP, MAXSAT, Operation Research (OR), SAT and QBF that indicate an improvement in the state-of-the-art solving heterogeneous instance sets. Last but not least, from our experimental studies, we deduce practical advice regarding the question when to apply which of our methods.
Surface displacement at volcanic edifices is related to subsurface processes associated with magma movements, fluid transfers within the volcano edifice and gravity-driven deformation processes. Understanding of associated ground displacements is of importance for assessment of volcanic hazards. For example, volcanic unrest is often preceded by surface uplift, caused by magma intrusion and followed by subsidence, after the withdrawal of magma. Continuous monitoring of the surface displacement at volcanoes therefore might allow the forecasting of upcoming eruptions to some extent. In geophysics, the measured surface displacements allow the parameters of possible deformation sources to be estimated through analytical or numerical modeling. This is one way to improve the understanding of subsurface processes acting at volcanoes. Although the monitoring of volcanoes has significantly improved in the last decades (in terms of technical advancements and number of monitored volcanoes), the forecasting of volcanic eruptions remains puzzling. In this work I contribute towards the understanding of the subsurface processes at volcanoes and thus to the improvement of volcano eruption forecasting. I have investigated the displacement field of Llaima volcano in Chile and of Tendürek volcano in East Turkey by using synthetic aperture radar interferometry (InSAR). Through modeling of the deformation sources with the extracted displacement data, it was possible to gain insights into potential subsurface processes occurring at these two volcanoes that had been barely studied before. The two volcanoes, although of very different origin, composition and geometry, both show a complexity of interacting deformation sources. At Llaima volcano, the InSAR technique was difficult to apply, due to the large decorrelation of the radar signal between the acquisition of images. I developed a model-based unwrapping scheme, which allows the production of reliable displacement maps at the volcano that I used for deformation source modeling. The modeling results show significant differences in pre- and post-eruptive magmatic deformation source parameters. Therefore, I conjecture that two magma chambers exist below Llaima volcano: a post-eruptive deep one and a shallow one possibly due to the pre-eruptive ascent of magma. Similar reservoir depths at Llaima have been confirmed by independent petrologic studies. These reservoirs are interpreted to be temporally coupled. At Tendürek volcano I have found long-term subsidence of the volcanic edifice, which can be described by a large, magmatic, sill-like source that is subject to cooling contraction. The displacement data in conjunction with high-resolution optical images, however, reveal arcuate fractures at the eastern and western flank of the volcano. These are most likely the surface expressions of concentric ring-faults around the volcanic edifice that show low magnitudes of slip over a long time. This might be an alternative mechanism for the development of large caldera structures, which are so far assumed to be generated during large catastrophic collapse events. To investigate the potential subsurface geometry and relation of the two proposed interacting sources at Tendürek, a sill-like magmatic source and ring-faults, I have performed a more sophisticated numerical modeling approach. The optimum source geometries show, that the size of the sill-like source was overestimated in the simple models and that it is difficult to determine the dip angle of the ring-faults with surface displacement data only. However, considering physical and geological criteria a combination of outward-dipping reverse faults in the west and inward-dipping normal faults in the east seem to be the most likely. Consequently, the underground structure at the Tendürek volcano consists of a small, sill-like, contracting, magmatic source below the western summit crater that causes a trapdoor-like faulting along the ring-faults around the volcanic edifice. Therefore, the magmatic source and the ring-faults are also interpreted to be temporally coupled. In addition, a method for data reduction has been improved. The modeling of subsurface deformation sources requires only a relatively small number of well distributed InSAR observations at the earth’s surface. Satellite radar images, however, consist of several millions of these observations. Therefore, the large amount of data needs to be reduced by several orders of magnitude for source modeling, to save computation time and increase model flexibility. I have introduced a model-based subsampling approach in particular for heterogeneously-distributed observations. It allows a fast calculation of the data error variance-covariance matrix, also supports the modeling of time dependent displacement data and is, therefore, an alternative to existing methods.
Metabolic systems tend to exhibit steady states that can be measured in terms of their concentrations and fluxes. These measurements can be regarded as a phenotypic representation of all the complex interactions and regulatory mechanisms taking place in the underlying metabolic network. Such interactions determine the system's response to external perturbations and are responsible, for example, for its asymptotic stability or for oscillatory trajectories around the steady state. However, determining these perturbation responses in the absence of fully specified kinetic models remains an important challenge of computational systems biology. Structural kinetic modeling (SKM) is a framework to analyse whether a metabolic steady state remains stable under perturbation, without requiring detailed knowledge about individual rate equations. It provides a parameterised representation of the system's Jacobian matrix in which the model parameters encode information about the enzyme-metabolite interactions. Stability criteria can be derived by generating a large number of structural kinetic models (SK-models) with randomly sampled parameter sets and evaluating the resulting Jacobian matrices. The parameter space can be analysed statistically in order to detect network positions that contribute significantly to the perturbation response. Because the sampled parameters are equivalent to the elasticities used in metabolic control analysis (MCA), the results are easy to interpret biologically. In this project, the SKM framework was extended by several novel methodological improvements. These improvements were evaluated in a simulation study using a set of small example pathways with simple Michaelis Menten rate laws. Afterwards, a detailed analysis of the dynamic properties of the neuronal TCA cycle was performed in order to demonstrate how the new insights obtained in this work could be used for the study of complex metabolic systems. The first improvement was achieved by examining the biological feasibility of the elasticity combinations created during Monte Carlo sampling. Using a set of small example systems, the findings showed that the majority of sampled SK-models would yield negative kinetic parameters if they were translated back into kinetic models. To overcome this problem, a simple criterion was formulated that mitigates such infeasible models and the application of this criterion changed the conclusions of the SKM experiment. The second improvement of this work was the application of supervised machine-learning approaches in order to analyse SKM experiments. So far, SKM experiments have focused on the detection of individual enzymes to identify single reactions important for maintaining the stability or oscillatory trajectories. In this work, this approach was extended by demonstrating how SKM enables the detection of ensembles of enzymes or metabolites that act together in an orchestrated manner to coordinate the pathways response to perturbations. In doing so, stable and unstable states served as class labels, and classifiers were trained to detect elasticity regions associated with stability and instability. Classification was performed using decision trees and relevance vector machines (RVMs). The decision trees produced good classification accuracy in terms of model bias and generalizability. RVMs outperformed decision trees when applied to small models, but encountered severe problems when applied to larger systems because of their high runtime requirements. The decision tree rulesets were analysed statistically and individually in order to explore the role of individual enzymes or metabolites in controlling the system's trajectories around steady states. The third improvement of this work was the establishment of a relationship between the SKM framework and the related field of MCA. In particular, it was shown how the sampled elasticities could be converted to flux control coefficients, which were then investigated for their predictive information content in classifier training. After evaluation on the small example pathways, the methodology was used to study two steady states of the neuronal TCA cycle with respect to their intrinsic mechanisms responsible for stability or instability. The findings showed that several elasticities were jointly coordinated to control stability and that the main source for potential instabilities were mutations in the enzyme alpha-ketoglutarate dehydrogenase.
Atmospheric interactions with land surface in the arctic based on regional climate model solutions
(2014)
Ausprägungen räumlicher Identität in ehemaligen sudetendeutschen Gebieten der Tschechischen Republik
(2014)
Das tschechische Grenzgebiet ist eine der Regionen in Europa, die in der Folge des Zweiten Weltkrieges am gravierendsten von Umbrüchen in der zuvor bestehenden Bevölkerungsstruktur betroffen waren. Der erzwungenen Aussiedlung eines Großteils der ansässigen Bevölkerung folgten die Neubesiedlung durch verschiedenste Zuwanderergruppen sowie teilweise langanhaltende Fluktuationen der Einwohnerschaft. Die Stabilisierung der Bevölkerung stand sodann unter dem Zeichen der sozialistischen Gesellschafts- und Wirtschaftsordnung, die die Lebensweise und Raumwahrnehmung der neuen Einwohner nachhaltig prägte. Die Grenzöffnung von 1989, die politische Transformation sowie die Integration der Tschechischen Republik in die Europäische Union brachten neue demographische und sozioökonomische Entwicklungen mit sich. Sie schufen aber auch die Bedingungen dafür, sich neu und offen auch mit der spezifischen Geschichte des ehemaligen Sudetenlandes sowie mit dem Zustand der gegenwärtigen Gesellschaft in diesem Gebiet auseinanderzusetzen.
Im Rahmen der vorliegenden Arbeit wird anhand zweier Beispielregionen untersucht, welche Raumvorstellungen und Raumbindungen bei der heute in den ehemaligen sudetendeutschen Gebieten ansässigen Bevölkerung vorhanden sind und welche Einflüsse die unterschiedlichen raumstrukturellen Bedingungen darauf ausüben. Besonderes Augenmerk wird auf die soziale Komponente der Ausprägung räumlicher Identität gelegt, das heißt auf die Rolle von Bedeutungszuweisungen gegenüber Raumelementen im Rahmen sozialer Kommunikation und Interaktion. Dies erscheint von besonderer Relevanz in einem Raum, der sich durch eine gewisse Heterogenität seiner Einwohnerschaft hinsichtlich ihres ethnischen, kulturellen beziehungsweise biographischen Hintergrundes auszeichnet. Schließlich wird ermittelt, welche Impulse unter Umständen von einer ausgeprägten räumlichen Identität für die Entwicklung des Raumes ausgehen.
Automated location of seismic events is a very important task in microseismic monitoring operations as well for local and regional seismic monitoring. Since microseismic records are generally characterised by low signal-to-noise ratio, such methods are requested to be noise robust and sufficiently accurate. Most of the standard automated location routines are based on the automated picking, identification and association of the first arrivals of P and S waves and on the minimization of the residuals between theoretical and observed arrival times of the considered seismic phases. Although current methods can accurately pick P onsets, the automatic picking of the S onset is still problematic, especially when the P coda overlaps the S wave onset. In this thesis I developed a picking free automated method based on the Short-Term-Average/Long-Term-Average (STA/LTA) traces at different stations as observed data. I used the STA/LTA of several characteristic functions in order to increase the sensitiveness to the P wave and the S waves. For the P phases we use the STA/LTA traces of the vertical energy function, while for the S phases, we use the STA/LTA traces of the horizontal energy trace and then a more optimized characteristic function which is obtained using the principal component analysis technique. The orientation of the horizontal components can be retrieved by robust and linear approach of waveform comparison between stations within a network using seismic sources outside the network (chapter 2). To locate the seismic event, we scan the space of possible hypocentral locations and origin times, and stack the STA/LTA traces along the theoretical arrival time surface for both P and S phases. Iterating this procedure on a three-dimensional grid we retrieve a multidimensional matrix whose absolute maximum corresponds to the spatial and temporal coordinates of the seismic event. Location uncertainties are then estimated by perturbing the STA/LTA parameters (i.e the length of both long and short time windows) and relocating each event several times. In order to test the location method I firstly applied it to a set of 200 synthetic events. Then we applied it to two different real datasets. A first one related to mining induced microseismicity in a coal mine in the northern Germany (chapter 3). In this case we successfully located 391 microseismic event with magnitude range between 0.5 and 2.0 Ml. To further validate the location method I compared the retrieved locations with those obtained by manual picking procedure. The second dataset consist in a pilot application performed in the Campania-Lucania region (southern Italy) using a 33 stations seismic network (Irpinia Seismic Network) with an aperture of about 150 km (chapter 4). We located 196 crustal earthquakes (depth < 20 km) with magnitude range 1.1 < Ml < 2.7. A subset of these locations were compared with accurate locations retrieved by a manual location procedure based on the use of a double difference technique. In both cases results indicate good agreement with manual locations. Moreover, the waveform stacking location method results noise robust and performs better than classical location methods based on the automatic picking of the P and S waves first arrivals.
In the field of disk-based parallel database management systems exists a great variety of solutions based on a shared-storage or a shared-nothing architecture. In contrast, main memory-based parallel database management systems are dominated solely by the shared-nothing approach as it preserves the in-memory performance advantage by processing data locally on each server. We argue that this unilateral development is going to cease due to the combination of the following three trends: a) Nowadays network technology features remote direct memory access (RDMA) and narrows the performance gap between accessing main memory inside a server and of a remote server to and even below a single order of magnitude. b) Modern storage systems scale gracefully, are elastic, and provide high-availability. c) A modern storage system such as Stanford's RAMCloud even keeps all data resident in main memory. Exploiting these characteristics in the context of a main-memory parallel database management system is desirable. The advent of RDMA-enabled network technology makes the creation of a parallel main memory DBMS based on a shared-storage approach feasible.
This thesis describes building a columnar database on shared main memory-based storage. The thesis discusses the resulting architecture (Part I), the implications on query processing (Part II), and presents an evaluation of the resulting solution in terms of performance, high-availability, and elasticity (Part III).
In our architecture, we use Stanford's RAMCloud as shared-storage, and the self-designed and developed in-memory AnalyticsDB as relational query processor on top. AnalyticsDB encapsulates data access and operator execution via an interface which allows seamless switching between local and remote main memory, while RAMCloud provides not only storage capacity, but also processing power. Combining both aspects allows pushing-down the execution of database operators into the storage system. We describe how the columnar data processed by AnalyticsDB is mapped to RAMCloud's key-value data model and how the performance advantages of columnar data storage can be preserved.
The combination of fast network technology and the possibility to execute database operators in the storage system opens the discussion for site selection. We construct a system model that allows the estimation of operator execution costs in terms of network transfer, data processed in memory, and wall time. This can be used for database operators that work on one relation at a time - such as a scan or materialize operation - to discuss the site selection problem (data pull vs. operator push). Since a database query translates to the execution of several database operators, it is possible that the optimal site selection varies per operator. For the execution of a database operator that works on two (or more) relations at a time, such as a join, the system model is enriched by additional factors such as the chosen algorithm (e.g. Grace- vs. Distributed Block Nested Loop Join vs. Cyclo-Join), the data partitioning of the respective relations, and their overlapping as well as the allowed resource allocation.
We present an evaluation on a cluster with 60 nodes where all nodes are connected via RDMA-enabled network equipment. We show that query processing performance is about 2.4x slower if everything is done via the data pull operator execution strategy (i.e. RAMCloud is being used only for data access) and about 27% slower if operator execution is also supported inside RAMCloud (in comparison to operating only on main memory inside a server without any network communication at all). The fast-crash recovery feature of RAMCloud can be leveraged to provide high-availability, e.g. a server crash during query execution only delays the query response for about one second. Our solution is elastic in a way that it can adapt to changing workloads a) within seconds, b) without interruption of the ongoing query processing, and c) without manual intervention.
This work introduces concepts and corresponding tool support to enable a complementary approach in dealing with recovery. Programmers need to recover a development state, or a part thereof, when previously made changes reveal undesired implications. However, when the need arises suddenly and unexpectedly, recovery often involves expensive and tedious work. To avoid tedious work, literature recommends keeping away from unexpected recovery demands by following a structured and disciplined approach, which consists of the application of various best practices including working only on one thing at a time, performing small steps, as well as making proper use of versioning and testing tools. However, the attempt to avoid unexpected recovery is both time-consuming and error-prone. On the one hand, it requires disproportionate effort to minimize the risk of unexpected situations. On the other hand, applying recommended practices selectively, which saves time, can hardly avoid recovery. In addition, the constant need for foresight and self-control has unfavorable implications. It is exhaustive and impedes creative problem solving. This work proposes to make recovery fast and easy and introduces corresponding support called CoExist. Such dedicated support turns situations of unanticipated recovery from tedious experiences into pleasant ones. It makes recovery fast and easy to accomplish, even if explicit commits are unavailable or tests have been ignored for some time. When mistakes and unexpected insights are no longer associated with tedious corrective actions, programmers are encouraged to change source code as a means to reason about it, as opposed to making changes only after structuring and evaluating them mentally. This work further reports on an implementation of the proposed tool support in the Squeak/Smalltalk development environment. The development of the tools has been accompanied by regular performance and usability tests. In addition, this work investigates whether the proposed tools affect programmers’ performance. In a controlled lab study, 22 participants improved the design of two different applications. Using a repeated measurement setup, the study examined the effect of providing CoExist on programming performance. The result of analyzing 88 hours of programming suggests that built-in recovery support as provided with CoExist positively has a positive effect on programming performance in explorative programming tasks.
Entrepreneurship is known to be a main driver of economic growth. Hence, governments have an interest in supporting and promoting entrepreneurial activities. Start-up subsidies, which have been analyzed extensively, only aim at mitigating the lack of financial capital. However, some entrepreneurs also lack in human, social, and managerial capital. One way to address these shortcomings is by subsidizing coaching programs for entrepreneurs. However, theoretical and empirical evidence about business coaching and programs subsidizing coaching is scarce. This dissertation gives an extensive overview of coaching and is the first empirical study for Germany analyzing the effects of coaching programs on its participants. In the theoretical part of the dissertation the process of a business start-up is described and it is discussed how and in which stage of the company’s evolvement coaching can influence entrepreneurial success. The concept of coaching is compared to other non-monetary types of support as training, mentoring, consulting, and counseling. Furthermore, national and international support programs are described. Most programs have either no or small positive effects. However, there is little quantitative evidence in the international literature. In the empirical part of the dissertation the effectiveness of coaching is shown by evaluating two German coaching programs, which support entrepreneurs via publicly subsidized coaching sessions. One of the programs aims at entrepreneurs who have been employed before becoming self-employed, whereas the other program is targeted at former unemployed entrepreneurs. The analysis is based on the evaluation of a quantitative and a qualitative dataset. The qualitative data are gathered by intensive one-on-one interviews with coaches and entrepreneurs. These data give a detailed insight about the coaching topics, duration, process, effectiveness, and the thoughts of coaches and entrepreneurs. The quantitative data include information about 2,936 German-based entrepreneurs. Using propensity score matching, the success of participants of the two coaching programs is compared with adequate groups of non-participants. In contrast to many other studies also personality traits are observed and controlled for in the matching process. The results show that only the program for former unemployed entrepreneurs has small positive effects. Participants have a larger survival probability in self-employment and a larger probability to hire employees than matched non-participants. In contrast, the program for former employed individuals has negative effects. Compared to individuals who did not participate in the coaching program, participants have a lower probability to stay in self-employment, lower earned net income, lower number of employees and lower life satisfaction. There are several reasons for these differing results of the two programs. First, former unemployed individuals have more basic coaching needs than former employed individuals. Coaches can satisfy these basic coaching needs, whereas former employed individuals have more complex business problems, which are not very easy to be solved by a coaching intervention. Second, the analysis reveals that former employed individuals are very successful in general. It is easier to increase the success of former unemployed individuals as they have a lower base level of success than former employed individuals. An effect heterogeneity analysis shows that coaching effectiveness differs by region. Coaching for previously unemployed entrepreneurs is especially useful in regions with bad labor market conditions. In summary, in line with previous literature, it is found that coaching has little effects on the success of entrepreneurs. The previous employment status, the characteristics of the entrepreneur and the regional labor market conditions play a crucial role in the effectiveness of coaching. In conclusion, coaching needs to be well tailored to the individual and applied thoroughly. Therefore, governments should design and provide coaching programs only after due consideration.
The contractile vacuole (CV) is an osmoregulatory organelle found exclusively in algae and protists. In addition to expelling excessive water out of the cell, it also expels ions and other metabolites and thereby contributes to the cell's metabolic homeostasis. The interest in the CV reaches beyond its immediate cellular roles. The CV's function is tightly related to basic cellular processes such as membrane dynamics and vesicle budding and fusion; several physiological processes in animals, such as synaptic neurotransmission and blood filtration in the kidney, are related to the CV's function; and several pathogens, such as the causative agents of sleeping sickness, possess CVs, which may serve as pharmacological targets. The green alga Chlamydomonas reinhardtii has two CVs. They are the smallest known CVs in nature, and they remain relatively untouched in the CV-related literature. Many genes that have been shown to be related to the CV in other organisms have close homologues in C. reinhardtii. We attempted to silence some of these genes and observe the effect on the CV. One of our genes, VMP1, caused striking, severe phenotypes when silenced. Cells exhibited defective cytokinesis and aberrant morphologies. The CV, incidentally, remained unscathed. In addition, mutant cells showed some evidence of disrupted autophagy. Several important regulators of the cell cycle as well as autophagy were found to be underexpressed in the mutant. Lipidomic analysis revealed many meaningful changes between wild-type and mutant cells, reinforcing the compromised-autophagy observation. VMP1 is a singular protein, with homologues in numerous eukaryotic organisms (aside from fungi), but usually with no relatives in each particular genome. Since its first characterization in 2002 it has been associated with several cellular processes and functions, namely autophagy, programmed cell-death, secretion, cell adhesion, and organelle biogenesis. It has been implicated in several human diseases: pancreatitis, diabetes, and several types of cancer. Our results reiterate some of the observations in VMP1's six reported homologues, but, importantly, show for the first time an involvement of this protein in cell division. The mechanisms underlying this involvement in Chlamydomonas, as well as other key aspects, such as VMP1's subcellular localization and interaction partners, still await elucidation.
Characterization of drought tolerance in potato cultivars for identification of molecular markers
(2014)
The adaptation of cell growth and proliferation to environmental changes is essential for the surviving of biological systems. The evolutionary conserved Ser/Thr protein kinase “Target of Rapamycin” (TOR) has emerged as a major signaling node that integrates the sensing of numerous growth signals to the coordinated regulation of cellular metabolism and growth. Although the TOR signaling pathway has been widely studied in heterotrophic organisms, the research on TOR in photosynthetic eukaryotes has been hampered by the reported land plant resistance to rapamycin. Thus, the finding that Chlamydomonas reinhardtii is sensitive to rapamycin, establish this unicellular green alga as a useful model system to investigate TOR signaling in photosynthetic eukaryotes.
The observation that rapamycin does not fully arrest Chlamydomonas growth, which is different from observations made in other organisms, prompted us to investigate the regulatory function of TOR in Chlamydomonas in context of the cell cycle. Therefore, a growth system that allowed synchronously growth under widely unperturbed cultivation in a fermenter system was set up and the synchronized cells were characterized in detail. In a highly resolved kinetic study, the synchronized cells were analyzed for their changes in cytological parameters as cell number and size distribution and their starch content. Furthermore, we applied mass spectrometric analysis for profiling of primary and lipid metabolism. This system was then used to analyze the response dynamics of the Chlamydomonas metabolome and lipidome to TOR-inhibition by rapamycin
The results show that TOR inhibition reduces cell growth, delays cell division and daughter cell release and results in a 50% reduced cell number at the end of the cell cycle. Consistent with the growth phenotype we observed strong changes in carbon and nitrogen partitioning in the direction of rapid conversion into carbon and nitrogen storage through an accumulation of starch, triacylglycerol and arginine. Interestingly, it seems that the conversion of carbon into triacylglycerol occurred faster than into starch after TOR inhibition, which may indicate a more dominant role of TOR in the regulation of TAG biosynthesis than in the regulation of starch.
This study clearly shows, for the first time, a complex picture of metabolic and lipidomic dynamically changes during the cell cycle of Chlamydomonas reinhardtii and furthermore reveals a complex regulation and adjustment of metabolite pools and lipid composition in response to TOR inhibition.
Cloud-RAID
(2014)
Synchronization is a fundamental phenomenon in nature. It can be considered as a general property of self-sustained oscillators to adjust their rhythm in the presence of an interaction.
In this work we investigate complex regimes of synchronization phenomena by means of theoretical analysis, numerical modeling, as well as practical analysis of experimental data.
As a subject of our investigation we consider chimera state, where due to spontaneous symmetry-breaking of an initially homogeneous oscillators lattice split the system into two parts with different dynamics. Chimera state as a new synchronization phenomenon was first found in non-locally coupled oscillators system, and has attracted a lot of attention in the last decade. However, the recent studies indicate that this state is also possible in globally coupled systems. In the first part of this work, we show under which conditions the chimera-like state appears in a system of globally coupled identical oscillators with intrinsic delayed feedback. The results of the research explain how initially monostable oscillators became effectivly bistable in the presence of the coupling and create a mean field that sustain the coexistence of synchronized and desynchronized states. Also we discuss other examples, where chimera-like state appears due to frequency dependence of the phase shift in the bistable system.
In the second part, we make further investigation of this topic by modeling influence of an external periodic force to an oscillator with intrinsic delayed feedback. We made stability analysis of the synchronized state and constructed Arnold tongues. The results explain formation of the chimera-like state and hysteric behavior of the synchronization area. Also, we consider two sets of parameters of the oscillator with symmetric and asymmetric Arnold tongues, that correspond to mono- and bi-stable regimes of the oscillator.
In the third part, we demonstrate the results of the work, which was done in collaboration with our colleagues from Psychology Department of University of Potsdam. The project aimed to study the effect of the cardiac rhythm on human perception of time using synchronization analysis. From our part, we made a statistical analysis of the data obtained from the conducted experiment on free time interval reproduction task. We examined how ones heartbeat influences the time perception and searched for possible phase synchronization between heartbeat cycles and time reproduction responses. The findings support the prediction that cardiac cycles can serve as input signals, and is used for reproduction of time intervals in the range of several seconds.
Previous studies on the acquisition of verb inflection in normally developing children have revealed an astonishing pattern: children use correctly inflected verbs in their own speech but fail to make use of verb inflections when comprehending sentences uttered by others. Thus, a three-year old might well be able to say something like ‘The cat sleeps on the bed’, but fails to understand that the same sentence, when uttered by another person, refers to only one sleeping cat but not more than one. The previous studies that have examined children's comprehension of verb inflections have employed a variant of a picture selection task in which the child was asked to explicitly indicate (via pointing) what semantic meaning she had inferred from the test sentence. Recent research on other linguistic structures, such as pronouns or focus particles, has indicated that earlier comprehension abilities can be found when methods are used that do not require an explicit reaction, like preferential looking tasks. This dissertation aimed to examine whether children are truly not able to understand the connection the the verb form and the meaning of the sentence subject until the age of five years or whether earlier comprehension can be found when a different measure, preferential looking, is used. Additionally, children's processing of subject-verb agreement violations was examined. The three experiments of this thesis that examined children's comprehension of verb inflections revealed the following: German-speaking three- to four-year old children looked more to a picture showing one actor when hearing a sentence with a singular inflected verb but only when their eye gaze was tracked and they did not have to perform a picture selection task. When they were asked to point to the matching picture, they performed at chance-level. This pattern indicates asymmetries in children's language performance even within the receptive modality. The fourth experiment examined sensitivity to subject-verb agreement violations and did not reveal evidence for sensitivity toward agreement violations in three- and four-year old children, but only found that children's looking patterns were influenced by the grammatical violations at the age of five. The results from these experiments are discussed in relation to the existence of a production-comprehension asymmetry in the use of verb inflections and children's underlying grammatical knowledge.
Despite remarkable progress made in the past century, which has revolutionized our understanding of the universe, there are numerous open questions left in theoretical physics. Particularly important is the fact that the theories describing the fundamental interactions of nature are incompatible. Einstein's theory of general relative describes gravity as a dynamical spacetime, which is curved by matter and whose curvature determines the motion of matter. On the other hand we have quantum field theory, in form of the standard model of particle physics, where particles interact via the remaining interactions - electromagnetic, weak and strong interaction - on a flat, static spacetime without gravity. A theory of quantum gravity is hoped to cure this incompatibility by heuristically replacing classical spacetime by quantum spacetime'. Several approaches exist attempting to define such a theory with differing underlying premises and ideas, where it is not clear which is to be preferred. Yet a minimal requirement is the compatibility with the classical theory, they attempt to generalize. Interestingly many of these models rely on discrete structures in their definition or postulate discreteness of spacetime to be fundamental. Besides the direct advantages discretisations provide, e.g. permitting numerical simulations, they come with serious caveats requiring thorough investigation: In general discretisations break fundamental diffeomorphism symmetry of gravity and are generically not unique. Both complicates establishing the connection to the classical continuum theory. The main focus of this thesis lies in the investigation of this relation for spin foam models. This is done on different levels of the discretisation / triangulation, ranging from few simplices up to the continuum limit. In the regime of very few simplices we confirm and deepen the connection of spin foam models to discrete gravity. Moreover, we discuss dynamical, e.g. diffeomorphism invariance in the discrete, to fix the ambiguities of the models. In order to satisfy these conditions, the discrete models have to be improved in a renormalisation procedure, which also allows us to study their continuum dynamics. Applied to simplified spin foam models, we uncover a rich, non--trivial fixed point structure, which we summarize in a phase diagram. Inspired by these methods, we propose a method to consistently construct the continuum theory, which comes with a unique vacuum state.
The background of civil service reform in Indonesia reveals the emergence of the reformation movement in 1998, following the fall of the authoritarian New Order regime. The reformation movement has seen the introduction of reforms in Indonesia's various governmental institutions, including the civil service. The civil service reforms were marked by the revision of Act 8/ 74 with Act 43 of 1999 on Civil Service Administration. The implementation of the civil service reform program, which was carried out by both central and local governments, required cooperation between the actors (in particular, Ministries, agencies and local governments), known as coordination.
Currently, the coordination that occurs between actors tends to be rigid and hierarchical. As a result, targets are not efficiently and effectively met. Hierarchical coordination, without a strong public sector infrastructure, tends to have a negative impact on achieving the desired outcomes of the civil service reform program. As an intrinsic part of the New Order regime, hierarchical coordination resulted in inefficiency and lack of efficacy. Despite these inefficiencies, the administration and the political environment have changed significantly as a result of the reform process.
Obvious examples of the reforms are changes in recruitment patterns, placement and remuneration policies. However, in the case of Indonesia, it appears that every state institution has its own policy. Thus, it appears that there has not been policy coherence in the civil service reform program, resulting in the lack of a sustainable program. The important thing to examine is how the coordination mechanisms of the civil service reform program in the central government have developed during the reform era in Indonesia
The purpose of this study is to analyse the linkages between coordination mechanisms and the actual implementation of civil service reform programs. This is undertaken as a basis for intervention based on the structures and patterns of coordination mechanisms in the implementation of civil service reform programs. The next step is to formulate the development coordination mechanisms, particularly to create structures and patterns of civil service reforms which are more sustainable to the specific characteristics of public sector organisations in Indonesia.
The benefits of this research are a stronger understanding of the linkages between the mechanisms of coordination and implementation of civil service reform programs. Through this analysis, the findings can then be applied as a basic consideration in planning a sustainable civil service reform program. In the basis of theoretical issues concerning the linkages between coordination mechanisms and implementation of civil service reform program sustainability, this book explores the type of coordination, which is needed to test the proportional and sustainable concept of the intended civil service reform program in Indonesia.
Research conducted through studies, surveys and donors has shown that poor coordination is the major hindrance to the civil service program reform in Indonesia. This research employs a qualitative approach. In this study, the coordination mechanisms and implementations of civil service reform programs are demonstrated by means of case studies of the State Ministry for Administrative Reform, the National Civil Service Agency and the National Public Administration Institute. The coordination mechanisms in these Ministries and agencies were analysed using indicators of effective and efficient coordination. The analysis of the coordination mechanisms shows a tendency towards rigid hierarchical coordination. This raises concerns about fragmentation among departments and agencies at the central government level and calls for integrated civil service reform both at central and local governmental levels. In the context of implementation programs, a hierarchical mechanism of coordination leverages on various aspects, such as the program formulation, implementation flow of the program, the impact of policies, and achievement targets. In particular, there was a shift process of the mainstream civil service reform in the Ministries and agencies which are marked by the emergence of sectoral interest and inefficiencies in the civil service reform program. The primary result of successful civil service reform is increased professionalism in the civil service.
The findings on hierarchical mechanisms and the prescriptions which will follow show that the professionalism of Indonesia’s civil service is at stake. The implementation of the program through coordination mechanisms in Ministries and agencies is measured in various dimensions: the centre of coordination, integration of coordination, sustainability of coordination and multidimensionality of coordination.
The results of this analysis show that coordination mechanisms and the implementation of civil service reform are more successful when they are integration rather than hierarchical mechanisms. For a successful implementation of the reform program, it is crucial to intervene and change the type of coordination at the central government through the integration approach (hierarchy, market, and network). Furthermore, in order to move towards the integration type mechanism of coordination the separation of the administration and politics in the practice of good governance needs to be carried out immediately and simultaneously. Based on this analysis, it can be concluded that the integration type mechanism of coordination is a suitable for Indonesia for a sustainable civil service reform program. Finally, to achieve coherent civil service reforms, national policies developed according to the central government's priorities are indispensable, establishing a coordination mechanism that can be adhered to throughout all reform sectors.
The H.E.S.S. array is a third generation Imaging Atmospheric Cherenkov Telescope (IACT) array. It is located in the Khomas Highland in Namibia, and measures very high energy (VHE) gamma-rays. In Phase I, the array started data taking in 2004 with its four identical 13 m telescopes. Since then, H.E.S.S. has emerged as the most successful IACT experiment to date. Among the almost 150 sources of VHE gamma-ray radiation found so far, even the oldest detection, the Crab Nebula, keeps surprising the scientific community with unexplained phenomena such as the recently discovered very energetic flares of high energy gamma-ray radiation. During its most recent flare, which was detected by the Fermi satellite in March 2013, the Crab Nebula was simultaneously observed with the H.E.S.S. array for six nights. The results of the observations will be discussed in detail during the course of this work. During the nights of the flare, the new 24 m × 32 m H.E.S.S. II telescope was still being commissioned, but participated in the data taking for one night. To be able to reconstruct and analyze the data of the H.E.S.S. Phase II array, the algorithms and software used by the H.E.S.S. Phase I array had to be adapted. The most prominent advanced shower reconstruction technique developed by de Naurois and Rolland, the template-based model analysis, compares real shower images taken by the Cherenkov telescope cameras with shower templates obtained using a semi-analytical model. To find the best fitting image, and, therefore, the relevant parameters that describe the air shower best, a pixel-wise log-likelihood fit is done. The adaptation of this advanced shower reconstruction technique to the heterogeneous H.E.S.S. Phase II array for stereo events (i.e. air showers seen by at least two telescopes of any kind), its performance using MonteCarlo simulations as well as its application to real data will be described.
The quantitative descriptions of the state of stress in the Earth’s crust, and spatial-temporal stress changes are of great importance in terms of scientific questions as well as applied geotechnical issues. Human activities in the underground (boreholes, tunnels, caverns, reservoir management, etc.) have a large impact on the stress state. It is important to assess, whether these activities may lead to (unpredictable) hazards, such as induced seismicity. Equally important is the understanding of the in situ stress state in the Earth’s crust, as it allows the determination of safe well paths, already during well planning. The same goes for the optimal configuration of the injection- and production wells, where stimulation for artificial fluid path ways is necessary.
The here presented cumulative dissertation consists of four separate manuscripts, which are already published, submitted or will be submitted for peer review within the next weeks. The main focus is on the investigation of the possible usage of geothermal energy in the province Alberta (Canada). A 3-D geomechanical–numerical model was designed to quantify the contemporary 3-D stress tensor in the upper crust. For the calibration of the regional model, 321 stress orientation data and 2714 stress magnitude data were collected, whereby the size and diversity of the database is unique. A calibration scheme was developed, where the model is calibrated versus the in situ stress data stepwise for each data type and gradually optimized using statistically test methods. The optimum displacement on the model boundaries can be determined by bivariate linear regression, based on only three model runs with varying deformation ratio. The best-fit model is able to predict most of the in situ stress data quite well. Thus, the model can provide the full stress tensor along any chosen virtual well paths. This can be used to optimize the orientation of horizontal wells, which e.g. can be used for reservoir stimulation. The model confirms regional deviations from the average stress orientation trend, such as in the region of the Peace River Arch and the Bow Island Arch.
In the context of data compilation for the Alberta stress model, the Canadian database of the World Stress Map (WSM) could be expanded by including 514 new data records. This publication of an update of the Canadian stress map after ~20 years with a specific focus on Alberta shows, that the maximum horizontal stress (SHmax) is oriented southwest to northeast over large areas in Northern America. The SHmax orientation in Alberta is very homogeneous, with an average of about 47°. In order to calculate the average SHmax orientation on a regular grid as well as to estimate the wave-length of stress orientation, an existing algorithm has been improved and is applied to the Canadian data. The newly introduced quasi interquartile range on the circle (QIROC) improves the variance estimation of periodic data, as it is less susceptible to its outliers.
Another geomechanical–numerical model was built to estimate the 3D stress tensor in the target area ”Nördlich Lägern” in Northern Switzerland. This location, with Opalinus clay as a host rock, is a potential repository site for high-level radioactive waste. The performed modelling aims to investigate the sensitivity of the stress tensor on tectonic shortening, topography, faults and variable rock properties within the Mesozoic sedimentary stack, according to the required stability needed for a suitable radioactive waste disposal site. The majority of the tectonic stresses caused by the far-field shortening from the South are admitted by the competent rock units in the footwall and hanging wall of the argillaceous target horizon, the Upper Malm and Upper Muschelkalk. Thus, the differential stress within the host rock remains relatively low. East-west striking faults release stresses driven by tectonic shortening. The purely gravitational influence by the topography is low; higher SHmax magnitudes below topographical depression and lower values below hills are mainly observed near the surface. A complete calibration of the model is not possible, as no stress magnitude data are available for calibration, yet. The collection of this data will begin in 2015; subsequently they will be used to adjust the geomechanical–numerical model again.
The third geomechanical–numerical model investigates the stress variation in an ultra-deep gold mine in South Africa. This reservoir model is spatially one order of magnitude smaller than the previous local model from Northern Switzerland. Here, the primary focus is to investigate the hypothesis that the Mw 1.9 earthquake on 27 December 2007 was induced by stress changes due to the mining process. The Coulomb failure stress change (DeltaCFS) was used to analyse the stress change. It confirmed that the seismic event was induced by static stress transfer due to the mining progress. The rock was brought closer to failure on the derived rupture plane by stress changes of up to 1.5–15MPa, in dependence of the DeltaCFS analysis type. A forward modelling of a generic excavation scheme reveals that with decreasing distance to the dyke the DeltaCFS values increase significantly. Hence, even small changes in the mining progress can have a significant impact on the seismic hazard risk, i.e. the change of the occurrence probability to induce a seismic event of economic concern.
In March 2010, the project CoCoCo (incipient COntinent-COntinent COllision) recorded a 650 km long amphibian N-S wide-angle seismic profile, extending from the Eratosthenes Seamount (ESM) across Cyprus and southern Turkey to the Anatolian plateau. The aim of the project is to reveal the impact of the transition from subduction to continent-continent collision of the African plate with the Cyprus-Anatolian plate. A visual quality check, frequency analysis and filtering were applied to the seismic data and reveal a good data quality. Subsequent first break picking, finite-differences ray tracing and inversion of the offshore wide-angle data leads to a first-arrival tomographic model. This model reveals (1) P-wave velocities lower than 6.5 km/s in the crust, (2) a variable crustal thickness of about 28 - 37 km and (3) an upper crustal reflection at 5 km depth beneath the ESM. Two land shots on Turkey, also recorded on Cyprus, airgun shots south of Cyprus and geological and previous seismic investigations provide the information to derive a layered velocity model beneath the Anatolian plateau and for the ophiolite complex on Cyprus. The analysis of the reflections provides evidence for a north-dipping plate subducting beneath Cyprus. The main features of this layered velocity model are (1) an upper and lower crust with large lateral changes of the velocity structure and thickness, (2) a Moho depth of about 38 - 45 km beneath the Anatolian plateau, (3) a shallow north-dipping subducting plate below Cyprus with an increasing dip and (4) a typical ophiolite sequence on Cyprus with a total thickness of about 12 km. The offshore-onshore seismic data complete and improve the information about the velocity structure beneath Cyprus and the deeper part of the offshore tomographic model. Thus, the wide-angle seismic data provide detailed insights into the 2-D geometry and velocity structures of the uplifted and overriding Cyprus-Anatolian plate. Subsequent gravity modelling confirms and extends the crustal P-wave velocity model. The deeper part of the subducting plate is constrained by the gravity data and has a dip angle of ~ 28°. Finally, an integrated analysis of the geophysical and geological information allows a comprehensive interpretation of the crustal structure related to the collision process.
Die Dissertation wird der rechtsstaatlichen Problematik in der Verfassung von Georgien und in der Rechtsprechung des Verfassungsgerichts Georgiens gewidmet und bietet eine umfangreiche Analyse der wichtigsten Rechtsstaatsmerkmale in dieser Hinsicht. Neben dieser Analyse der früheren und bestehenden Gegebenheiten, prognostiziert sie mittels der perspektivistischen Denkmodele zukünftige Entwicklungen und schlägt eventuelle Lösungswege von Problemen vor. Bei der Arbeit an der Dissertation wurde insbesondere die reiche Erfahrung der deutschen Rechtsstaatsdogmatik in Anspruch genommen.
The data quality of real-world datasets need to be constantly monitored and maintained to allow organizations and individuals to reliably use their data. Especially, data integration projects suffer from poor initial data quality and as a consequence consume more effort and money. Commercial products and research prototypes for data cleansing and integration help users to improve the quality of individual and combined datasets. They can be divided into either standalone systems or database management system (DBMS) extensions. On the one hand, standalone systems do not interact well with DBMS and require time-consuming data imports and exports. On the other hand, DBMS extensions are often limited by the underlying system and do not cover the full set of data cleansing and integration tasks.
We overcome both limitations by implementing a concise set of five data cleansing and integration operators on the parallel data analytics platform Stratosphere. We define the semantics of the operators, present their parallel implementation, and devise optimization techniques for individual operators and combinations thereof. Users specify declarative queries in our query language METEOR with our new operators to improve the data quality of individual datasets or integrate them to larger datasets. By integrating the data cleansing operators into the higher level language layer of Stratosphere, users can easily combine cleansing operators with operators from other domains, such as information extraction, to complex data flows. Through a generic description of the operators, the Stratosphere optimizer reorders operators even from different domains to find better query plans.
As a case study, we reimplemented a part of the large Open Government Data integration project GovWILD with our new operators and show that our queries run significantly faster than the original GovWILD queries, which rely on relational operators. Evaluation reveals that our operators exhibit good scalability on up to 100 cores, so that even larger inputs can be efficiently processed by scaling out to more machines. Finally, our scripts are considerably shorter than the original GovWILD scripts, which results in better maintainability of the scripts.
Der Häftlingsfreikauf aus der DDR 1962/63 - 1989 : zwischen Menschenhandel und humanitären Aktionen
(2014)
Schlucken ist ein lebensnotwendiger Prozess, dessen Diagnose und Therapie eine enorme Herausforderung bedeutet. Die Erkennung und Beurteilung von Schlucken und Schluckstörungen erfordert den Einsatz von technisch aufwendigen Verfahren, wie Videofluoroskopie (VFSS) und fiberoptisch-endoskopische Schluckuntersuchung (FEES), die eine hohe Belastung für die Patienten darstellen. Beide Verfahren werden als Goldstandard in der Diagnostik von Schluckstörungen eingesetzt. Die Durchführung obliegt in der Regel ärztlichem Personal. Darüber hinaus erfordert die Auswertung des Bildmaterials der Diagnostik eine ausreichend hohe Erfahrung. In der Therapie findet neben den klassischen Therapiemethoden, wie z.B. diätetische Modifikationen und Schluckmanöver, auch zunehmend die funktionelle Elektrostimulation Anwendung. Ziel der vorliegenden Dissertationsschrift ist die Evaluation eines im Verbundprojekt BigDysPro entwickelten Bioimpedanz (BI)- und Elektromyographie (EMG)-Messsystems. Es wurde geprüft, ob sich das BI- und EMG-Messsystem eignet, sowohl in der Diagnostik als auch in der Therapie als eigenständiges Messsystem und im Rahmen einer Schluckneuroprothese eingesetzt zu werden. In verschiedenen Studien wurden gesunde Probanden für die Überprüfung der Reproduzierbarkeit (Intra-und Interrater-Reliabilität), der Unterscheidbarkeit von Schluck- und Kopfbewegungen und der Beeinflussung der Biosignale (BI, EMG) durch verschiedene Faktoren (Geschlecht der Probanden, Leitfähigkeit, Konsistenz und Menge der Nahrung) untersucht. Durch zusätzliche Untersuchungen mit Patienten wurde einerseits der Einfluss der Elektrodenart geprüft. Andererseits wurden parallel zur BI- und EMG-Messung auch endoskopische (FEES) und radiologische Schluckuntersuchungen (VFSS) durchgeführt, um die Korrelation der Biosignale mit der Bewegung anatomischer Strukturen (VFSS) und mit der Schluckqualität (FEES) zu prüfen. Es wurden 31 gesunde Probanden mit 1819 Schlucken und 60 Patienten mit 715 Schlucken untersucht. Die Messkurven zeigten einen typischen, reproduzierbaren Signalverlauf, der mit anatomischen und funktionellen Änderungen während der pharyngalen Schluckphase in der VFSS korrelierte (r > 0,7). Aus dem Bioimpedanzsignal konnten Merkmale extrahiert werden, die mit physiologischen Merkmalen eines Schluckes, wie verzögerter laryngealer Verschluss und Kehlkopfhebung, korrelierten und eine Einschätzung der Schluckqualität in Übereinstimmung mit der FEES ermöglichten. In den Signalverläufen der Biosignale konnten signifikante Unterschiede zwischen Schluck- und Kopfbewegungen und den Nahrungsmengen und -konsistenzen nachgewiesen werden. Im Gegensatz zur Nahrungsmenge und -konsistenz zeigte die Leitfähigkeit der zu schluckenden Nahrung, das Geschlecht der Probanden und die Art der Elektroden keinen signifikanten Einfluss auf die Messsignale. Mit den Ergebnissen der Evaluation konnte gezeigt werden, dass mit dem BI- und EMG-Messsystem ein neuartiges und nicht-invasives Verfahren zur Verfügung steht, das eine reproduzierbare Darstellung der pharyngalen Schluckphase und ihrer Veränderungen ermöglicht. Daraus ergeben sich vielseitige Einsatzmöglichkeiten in der Diagnostik, z.B. Langzeitmessung zur Schluckfrequenz und Einschätzung der Schluckqualität, und in der Therapie, z.B. der Einsatz in einer Schluckneuroprothese oder als Biofeedback zur Darstellung des Schluckes, von Schluckstörungen.
Inhalt der Arbeit ist es, einen Überblick über die historische Entwicklung der öffentlich-rechtlichen Gefährdungshaftung in Deutschland vom 18. Jahrhundert bis heute zu geben sowie ihre praktische Bedeutung zu analysieren. Dabei wird zwischen den unterschiedlichen Gesetzgebungen, Rechtsprechungen und den theoretischen Lösungsansätzen der öffentlich-rechtlichen Gefährdungshaftung unterschieden und insbesondere letzteres problematisiert. Ferner wird auf das Verhältnis zu den grundrechtlichen Schutzpflichten, den sozialen Risikotatbeständen, dem sozialrechtlichen Herstellungsanspruch und den Tumultschäden eingegangen.
Durch Zeit und Raum
(2014)
Die vorliegende Studie beschäftigte sich mit der Bedeutung der dysfunktionalen Einstellungen für die Entwicklung von depressiven Symptomen bei Kindern und Jugendlichen. Nach der kognitiven Theorie der Depression von Beck (1967, 1996) führen dysfunktionale Einstellungen in Interaktion mit Stress zu depressiven Symptomen. Es existieren allerdings nur wenige Studien, die die longitudinale Beziehung zwischen den dysfunktionalen Einstellungen und der Depressivität bei Kindern und Jugendlichen untersucht haben (Lakdawalla et al., 2007). Folglich kann noch nicht eindeutig geklärt werden, ob die dysfunktionalen Einstellungen Ursache, Begleiterscheinung oder Konsequenz der Depression sind. Als Datengrundlage diente eine Stichprobe von Kindern und Jugendlichen im Alter von 9 bis 20 Jahren, die im Rahmen der PIER-Studie zu dysfunktionalen Einstellungen, kritischen Lebensereignissen und depressiven Symptomen befragt wurden (Nt1t2 = 1.053; t1: 2011/2012, t2: 2013/2014). Querschnittliche Analysen zeigten hohe Assoziationen zwischen den dysfunktionalen Einstellungen, kritischen Lebensereignissen und depressiven Symptomen. Eine latente Moderationsanalyse wies nur bei den Jugendlichen auf signifikante Interaktion zwischen den dysfunktionalen Einstellungen und den kritischen Lebensereignissen in der Vorhersage depressiver Symptomatik hin. Im Längsschnitt zeigten latente Cross-Lagged-Panel-Analysen erwartungsgemäß, dass die dysfunktionalen Einstellungen und die Depressivität mit dem Alter immer stabilere Konstrukte darstellen, die sehr eng miteinander zusammenhängen. Eine diesem Modell hinzugefügte latente Moderationsanalyse konnte das kognitive Modell der Depression nach Beck weder bei Kindern noch bei Jugendlichen bestätigen. Die spätere depressive Symptomatik konnte lediglich durch Haupteffekte der früheren Ausprägung der Depressivität und der kritischen Lebensereignisse vorhergesagt werden. Diese Ergebnisse legen den Schluss nahe, dass es sich bei den dysfunktionalen Einstellungen eher um Begleiterscheinungen als um Risikofaktoren oder Konsequenzen der depressiven Symptomatik handelt.
Effect of benzylglucosinolate on signaling pathways associated with type 2 diabetes prevention
(2014)
Type 2 diabetes (T2D) is a health problem throughout the world. In 2010, there were nearly 230 million individuals with diabetes worldwide and it is estimated that in the economically advanced countries the cases will increase about 50% in the next twenty years. Insulin resistance is one of major features in T2D, which is also a risk factor for metabolic and cardiovascular complications. Epidemiological and animal studies have shown that the consumption of vegetables and fruits can delay or prevent the development of the disease, although the underlying mechanisms of these effects are still unclear. Brassica species such as broccoli (Brassica oleracea var. italica) and nasturtium (Tropaeolum majus) possess high content of bioactive phytochemicals, e.g. nitrogen sulfur compounds (glucosinolates and isothiocyanates) and polyphenols largely associated with the prevention of cancer. Isothiocyanates (ITCs) display their anti-carcinogenic potential by inducing detoxicating phase II enzymes and increasing glutathione (GSH) levels in tissues. In T2D diabetes an increase in gluconeogenesis and triglyceride synthesis, and a reduction in fatty acid oxidation accompanied by the presence of reactive oxygen species (ROS) are observed; altogether is the result of an inappropriate response to insulin. Forkhead box O (FOXO) transcription factors play a crucial role in the regulation of insulin effects on gene expression and metabolism, and alterations in FOXO function could contribute to metabolic disorders in diabetes. In this study using stably transfected human osteosarcoma cells (U-2 OS) with constitutive expression of FOXO1 protein labeled with GFP (green fluorescent protein) and human hepatoma cells HepG2 cell cultures, the ability of benzylisothiocyanate (BITC) deriving from benzylglucosinolate, extracted from nasturtium to modulate, i) the insulin-signaling pathway, ii) the intracellular localization of FOXO1 and iii) the expression of proteins involved in glucose metabolism, ROS detoxification, cell cycle arrest and DNA repair was evaluated. BITC promoted oxidative stress and in response to that induced FOXO1 translocation from cytoplasm into the nucleus antagonizing the insulin effect. BITC stimulus was able to down-regulate gluconeogenic enzymes, which can be considered as an anti-diabetic effect; to promote antioxidant resistance expressed by the up-regulation in manganese superoxide dismutase (MnSOD) and detoxification enzymes; to modulate autophagy by induction of BECLIN1 and down-regulation of the mammalian target of rapamycin complex 1 (mTORC1) pathway; and to promote cell cycle arrest and DNA damage repair by up-regulation of the cyclin-dependent kinase inhibitor (p21CIP) and Growth Arrest / DNA Damage Repair (GADD45). Except for the nuclear factor (erythroid derived)-like2 (NRF2) and its influence in the detoxification enzymes gene expression, all the observed effects were independent from FOXO1, protein kinase B (AKT/PKB) and NAD-dependent deacetylase sirtuin-1 (SIRT1). The current study provides evidence that besides of the anticarcinogenic potential, isothiocyanates might have a role in T2D prevention. BITC stimulus mimics the fasting state, in which insulin signaling is not triggered and FOXO proteins remain in the nucleus modulating gene expression of their target genes, with the advantage of a down-regulation of gluconeogenesis instead of its increase. These effects suggest that BITC might be considered as a promising substance in the prevention or treatment of T2D, therefore the factors behind of its modulatory effects need further investigation.
Die Häufung von Diabetes, Herz-Kreislauf-Erkrankungen und einigen Krebsarten, deren Entstehung auf Übergewicht und Bewegungsmangel zurückzuführen sind, ist ein aktuelles Problem unserer Gesellschaft. Insbesondere mit fortschreitendem Alter nehmen die damit einhergehenden Komplikationen zu. Umso bedeutender ist das Verständnis der pathologischen Mechanismen in Folge von Adipositas, Bewegungsmangel, des Alterungsprozesses und den Einfluss-nehmenden Faktoren.
Ziel dieser Arbeit war die Entstehung metabolischer Erkrankungen beim Menschen zu untersuchen. Die Auswertung von Verlaufsdaten anthropometrischer und metabolischer Parameter der 584 Teilnehmern der prospektiven ‚Metabolisches Syndrom Berlin Potsdam Follow-up Studie‘ wies für die gesamte Kohorte einen Anstieg an Übergewicht, ebenso eine Verschlechterung des Blutdrucks und des Glukosestoffwechsels auf. Wir untersuchten, ob das Hormon FGF21 Einfluss an dem Auftreten eines Diabetes mellitus Typ 2 (T2DM) oder des Metabolischen Syndroms (MetS) hat. Wir konnten zeigen, dass Personen, die später ein MetS entwickeln, bereits zu Studienbeginn einen erhöhten FGF21-Spiegel, einen höheren BMI, WHR, Hb1Ac und diastolischen Blutdruck aufwiesen. Neben FGF21 wurde auch Vaspin in diesem Zusammenhang untersucht. Es zeigte sich, dass Personen, die später einen T2DM entwickeln, neben einer Erhöhung klinischer Parameter tendenziell erhöhte Spiegel des Hormons aufwiesen. Mit FGF21 und Vaspin wurden hier zwei neue Faktoren für die Vorhersage des Metabolischen Syndroms bzw. Diabetes mellitus Typ 2 identifiziert.
Der langfristige Effekt einer Gewichtsreduktion wurde in einer Subkohorte von 60 Personen untersucht. Der überwiegende Teil der Probanden mit Gewichtsabnahme-Intervention nahm in der ersten sechsmonatigen Phase erfolgreich ab. Jedoch zeigte sich ein deutlicher Trend zur Wiederzunahme des verlorenen Gewichts über den Beobachtungszeitraum von fünf Jahren. Von besonderem Interesse war die Abschätzung des kardiovaskulären Risikos über den Framingham Score. Es wurde deutlich, dass für Personen mit konstanter Gewichtsabnahme ein deutlich geringeres kardiovaskuläres Risiko bestand. Hingegen zeigten Personen mit konstanter Wiederzunahme oder starken Gewichtsschwankungen ein hohes kardiovaskuläres Risiko. Unsere Daten legten nahe, dass eine erfolgreiche dauerhafte Gewichtsreduktion statistisch mit einem erniedrigten kardiovaskulären Risiko assoziiert ist, während Probanden mit starken Gewichtsschwankungen oder einer Gewichtszunahme ein gesteigertes Risiko haben könnten.
Um die Interaktion der molekularen Vorgänge hinsichtlich der Gewichtsreduktion und Lebensspanne untersuchen zu können, nutzen wir den Modellorganismus C.elegans. Eine kontinuierliche Restriktion wirkte sich verlängernd, eine Überversorgung verkürzend auf die Lebensspanne des Rundwurms aus. Der Einfluss eines zeitlich eingeschränkten, intermittierenden Nahrungsregimes, analog zum Weight-Cycling im Menschen, auf die Lebensspanne war von großem Interesse. Dieser regelmäßige Wechsel zwischen ad libitum Fütterung und Restriktion hatte in Abhängigkeit von der Häufigkeit der Restriktion einen unterschiedlich starken lebensverlängernden Effekt. Phänomene, wie Gewichtswiederzunahmen, sind in C.elegans nicht zu beobachten und beruhen vermutlich auf einem Mechanismus ist, der evolutionär jünger und in C.elegans noch nicht angelegt ist.
Um neue Stoffwechselwege zu identifizieren, die die Lebensspanne beeinflussen, wurden Metabolitenprofile genetischer als auch diätetischer Langlebigkeitsmodelle analysiert. Diese Analysen wiesen den Tryptophan-Stoffwechsel als einen neuen, bisher noch nicht im Fokus stehenden Stoffwechselweg aus, der mit Langlebigkeit in Verbindung steht.
El lunfardo : Kontaktvarietät der Migrationskultur am Rio de la Plata und in der Welt des Tango
(2014)
Donor-acceptor (D-A) copolymers have revolutionized the field of organic electronics over the last decade. Comprised of a electron rich and an electron deficient molecular unit, these copolymers facilitate the systematic modification of the material's optoelectronic properties. The ability to tune the optical band gap and to optimize the molecular frontier orbitals as well as the manifold of structural sites that enable chemical modifications has created a tremendous variety of copolymer structures. Today, these materials reach or even exceed the performance of amorphous inorganic semiconductors. Most impressively, the charge carrier mobility of D-A copolymers has been pushed to the technologically important value of 10 cm^{2}V^{-1}s^{-1}. Furthermore, owed to their enormous variability they are the material of choice for the donor component in organic solar cells, which have recently surpassed the efficiency threshold of 10%. Because of the great number of available D-A copolymers and due to their fast chemical evolution, there is a significant lack of understanding of the fundamental physical properties of these materials. Furthermore, the complex chemical and electronic structure of D-A copolymers in combination with their semi-crystalline morphology impede a straightforward identification of the microscopic origin of their superior performance. In this thesis, two aspects of prototype D-A copolymers were analysed. These are the investigation of electron transport in several copolymers and the application of low band gap copolymers as acceptor component in organic solar cells. In the first part, the investigation of a series of chemically modified fluorene-based copolymers is presented. The charge carrier mobility varies strongly between the different derivatives, although only moderate structural changes on the copolymers structure were made. Furthermore, rather unusual photocurrent transients were observed for one of the copolymers. Numerical simulations of the experimental results reveal that this behavior arises from a severe trapping of electrons in an exponential distribution of trap states. Based on the comparison of simulation and experiment, the general impact of charge carrier trapping on the shape of photo-CELIV and time-of-flight transients is discussed. In addition, the high performance naphthalenediimide (NDI)-based copolymer P(NDI2OD-T2) was characterized. It is shown that the copolymer posses one of the highest electron mobilities reported so far, which makes it attractive to be used as the electron accepting component in organic photovoltaic cells.\par Solar cells were prepared from two NDI-containing copolymers, blended with the hole transporting polymer P3HT. I demonstrate that the use of appropriate, high boiling point solvents can significantly increase the power conversion efficiency of these devices. Spectroscopic studies reveal that the pre-aggregation of the copolymers is suppressed in these solvents, which has a strong impact on the blend morphology. Finally, a systematic study of P3HT:P(NDI2OD-T2) blends is presented, which quantifies the processes that limit the efficiency of devices. The major loss channel for excited states was determined by transient and steady state spectroscopic investigations: the majority of initially generated electron-hole pairs is annihilated by an ultrafast geminate recombination process. Furthermore, exciton self-trapping in P(NDI2OD-T2) domains account for an additional reduction of the efficiency. The correlation of the photocurrent to microscopic morphology parameters was used to disclose the factors that limit the charge generation efficiency. Our results suggest that the orientation of the donor and acceptor crystallites relative to each other represents the main factor that determines the free charge carrier yield in this material system. This provides an explanation for the overall low efficiencies that are generally observed in all-polymer solar cells.
In this thesis we consider diverse aspects of existence and correctness of asymptotic solutions to elliptic differential and pseudodifferential equations. We begin our studies with the case of a general elliptic boundary value problem in partial derivatives. A small parameter enters the coefficients of the main equation as well as into the boundary conditions. Such equations have already been investigated satisfactory, but there still exist certain theoretical deficiencies. Our aim is to present the general theory of elliptic problems with a small parameter. For this purpose we examine in detail the case of a bounded domain with a smooth boundary. First of all, we construct formal solutions as power series in the small parameter. Then we examine their asymptotic properties. It suffices to carry out sharp two-sided \emph{a priori} estimates for the operators of boundary value problems which are uniform in the small parameter. Such estimates failed to hold in functional spaces used in classical elliptic theory. To circumvent this limitation we exploit norms depending on the small parameter for the functions defined on a bounded domain. Similar norms are widely used in literature, but their properties have not been investigated extensively. Our theoretical investigation shows that the usual elliptic technique can be correctly carried out in these norms. The obtained results also allow one to extend the norms to compact manifolds with boundaries. We complete our investigation by formulating algebraic conditions on the operators and showing their equivalence to the existence of a priori estimates. In the second step, we extend the concept of ellipticity with a small parameter to more general classes of operators. Firstly, we want to compare the difference in asymptotic patterns between the obtained series and expansions for similar differential problems. Therefore we investigate the heat equation in a bounded domain with a small parameter near the time derivative. In this case the characteristics touch the boundary at a finite number of points. It is known that the solutions are not regular in a neighbourhood of such points in advance. We suppose moreover that the boundary at such points can be non-smooth but have cuspidal singularities. We find a formal asymptotic expansion and show that when a set of parameters comes through a threshold value, the expansions fail to be asymptotic. The last part of the work is devoted to general concept of ellipticity with a small parameter. Several theoretical extensions to pseudodifferential operators have already been suggested in previous studies. As a new contribution we involve the analysis on manifolds with edge singularities which allows us to consider wider classes of perturbed elliptic operators. We examine that introduced classes possess a priori estimates of elliptic type. As a further application we demonstrate how developed tools can be used to reduce singularly perturbed problems to regular ones.
In Chapter 1 of the dissertation, the role of social networks is analyzed as an important determinant in the search behavior of the unemployed. Based on the hypothesis that the unemployed generate information on vacancies through their social network, search theory predicts that individuals with large social networks should experience an increased productivity of informal search, and reduce their search in formal channels. Due to the higher productivity of search, unemployed with a larger network are also expected to have a higher reservation wage than unemployed with a small network. The model-theoretic predictions are tested and confirmed empirically. It is found that the search behavior of unemployed is significantly affected by the presence of social contacts, with larger networks implying a stronger substitution away from formal search channels towards informal channels. The substitution is particularly pronounced for passive formal search methods, i.e., search methods that generate rather non-specific types of job offer information at low relative cost. We also find small but significant positive effects of an increase of the network size on the reservation wage. These results have important implications on the analysis of the job search monitoring or counseling measures that are usually targeted at formal search only. Chapter 2 of the dissertation addresses the labor market effects of vacancy information during the early stages of unemployment. The outcomes considered are the speed of exit from unemployment, the effects on the quality of employment and the short-and medium-term effects on active labor market program (ALMP) participation. It is found that vacancy information significantly increases the speed of entry into employment; at the same time the probability to participate in ALMP is significantly reduced. Whereas the long-term reduction in the ALMP arises in consequence of the earlier exit from unemployment, we also observe a short-run decrease for some labor market groups which suggest that caseworker use high and low intensity activation measures interchangeably which is clearly questionable from an efficiency point of view. For unemployed who find a job through vacancy information we observe a small negative effect on the weekly number of hours worked. In Chapter 3, the long-term effects of participation in ALMP are assessed for unemployed youth under 25 years of age. Complementary to the analysis in Chapter 2, the effects of participation in time- and cost-intensive measures of active labor market policies are examined. In particular we study the effects of job creation schemes, wage subsidies, short-and long-term training measures and measures to promote the participation in vocational training. The outcome variables of interest are the probability to be in regular employment, and participation in further education during the 60 months following program entry. The analysis shows that all programs, except job creation schemes have positive and long-term effects on the employment probability of youth. In the short-run only short-term training measures generate positive effects, as long-term training programs and wage subsidies exhibit significant locking-in'' effects. Measures to promote vocational training are found to increase the probability of attending education and training significantly, whereas all other programs have either no or a negative effect on training participation. Effect heterogeneity with respect to the pre-treatment level education shows that young people with higher pre-treatment educational levels benefit more from participation most programs. However, for longer-term wage subsidies we also find strong positive effects for young people with low initial education levels. The relative benefit of training measures is higher in West than in East Germany. In the evaluation studies of Chapters 2 and 3 semi-parametric balancing methods of Propensity Score Matching (PSM) and Inverse Probability Weighting (IPW) are used to eliminate the effects of counfounding factors that influence both the treatment participation as well as the outcome variable of interest, and to establish a causal relation between program participation and outcome differences. While PSM and IPW are intuitive and methodologically attractive as they do not require parametric assumptions, the practical implementation may become quite challenging due to their sensitivity to various data features. Given the importance of these methods in the evaluation literature, and the vast number of recent methodological contributions in this field, Chapter 4 aims to reduce the knowledge gap between the methodological and applied literature by summarizing new findings of the empirical and statistical literature and practical guidelines for future applied research. In contrast to previous publications this study does not only focus on the estimation of causal effects, but stresses that the balancing challenge can and should be discussed independent of question of causal identification of treatment effects on most empirical applications. Following a brief outline of the practical implementation steps required for PSM and IPW, these steps are presented in detail chronologically, outlining practical advice for each step. Subsequently, the topics of effect estimation, inference, sensitivity analysis and the combination with parametric estimation methods are discussed. Finally, new extensions of the methodology and avenues for future research are presented.
Virtualized cloud data centers provide on-demand resources, enable agile resource provisioning, and host heterogeneous applications with different resource requirements. These data centers consume enormous amounts of energy, increasing operational expenses, inducing high thermal inside data centers, and raising carbon dioxide emissions. The increase in energy consumption can result from ineffective resource management that causes inefficient resource utilization. This dissertation presents detailed models and novel techniques and algorithms for virtual resource management in cloud data centers. The proposed techniques take into account Service Level Agreements (SLAs) and workload heterogeneity in terms of memory access demand and communication patterns of web applications and High Performance Computing (HPC) applications. To evaluate our proposed techniques, we use simulation and real workload traces of web applications and HPC applications and compare our techniques against the other recently proposed techniques using several performance metrics. The major contributions of this dissertation are the following: proactive resource provisioning technique based on robust optimization to increase the hosts' availability for hosting new VMs while minimizing the idle energy consumption. Additionally, this technique mitigates undesirable changes in the power state of the hosts by which the hosts' reliability can be enhanced in avoiding failure during a power state change. The proposed technique exploits the range-based prediction algorithm for implementing robust optimization, taking into consideration the uncertainty of demand. An adaptive range-based prediction for predicting workload with high fluctuations in the short-term. The range prediction is implemented in two ways: standard deviation and median absolute deviation. The range is changed based on an adaptive confidence window to cope with the workload fluctuations. A robust VM consolidation for efficient energy and performance management to achieve equilibrium between energy and performance trade-offs. Our technique reduces the number of VM migrations compared to recently proposed techniques. This also contributes to a reduction in energy consumption by the network infrastructure. Additionally, our technique reduces SLA violations and the number of power state changes. A generic model for the network of a data center to simulate the communication delay and its impact on VM performance, as well as network energy consumption. In addition, a generic model for a memory-bus of a server, including latency and energy consumption models for different memory frequencies. This allows simulating the memory delay and its influence on VM performance, as well as memory energy consumption. Communication-aware and energy-efficient consolidation for parallel applications to enable the dynamic discovery of communication patterns and reschedule VMs using migration based on the determined communication patterns. A novel dynamic pattern discovery technique is implemented, based on signal processing of network utilization of VMs instead of using the information from the hosts' virtual switches or initiation from VMs. The result shows that our proposed approach reduces the network's average utilization, achieves energy savings due to reducing the number of active switches, and provides better VM performance compared to CPU-based placement. Memory-aware VM consolidation for independent VMs, which exploits the diversity of VMs' memory access to balance memory-bus utilization of hosts. The proposed technique, Memory-bus Load Balancing (MLB), reactively redistributes VMs according to their utilization of a memory-bus using VM migration to improve the performance of the overall system. Furthermore, Dynamic Voltage and Frequency Scaling (DVFS) of the memory and the proposed MLB technique are combined to achieve better energy savings.
Enterprise-specific in-memory data managment : HYRISEc - an in-memory column store engine for OLXP
(2014)
Weltweit streben Anti-Doping Institute danach jene Sportler zu überführen, welche sich unerlaubter Mittel oder Methoden bedienen. Die hierfür notwendigen Testsysteme werden kontinuierlich weiterentwickelt und neue Methoden aufgrund neuer Wirkstoffe der Pharmaindustrie etabliert. Gegenstand dieser Arbeit war es, eine parallele Mehrkomponentenanalyse auf Basis von Antigen-Antikörper Reaktionen zu entwickeln, bei dem es primär um Verringerung des benötigten Probevolumens und der Versuchszeit im Vergleich zu einem Standard Nachweis-Verfahren ging. Neben der Verwendung eines Multiplex Ansatzes und der Mikroarraytechnologie stellten ebenfalls die Genauigkeit aller Messparameter, die Stabilität des Versuchsaufbaus sowie die Performance über einen Einfach-Blind-Ansatz Herausforderungen dar. Die Anforderung an den Multiplex Ansatz, keine falschen Signale trotz ähnlicher Strukturen zu messen, konnte durch die gezielte Kombination von spezifischen Antikörpern realisiert werden. Hierfür wurden neben Kreuzreaktivitätstests auf dem Mikroarray parallel erfolgreich Western Blot Versuche durchgeführt. Jene Antikörper, welche in diesen Versuchen die gesetzten Anforderungen erfüllten, wurden für das Ermitteln der kleinsten nachweisbaren Konzentration verwendet. Über das Optimieren der Versuchsbedingungen konnte unter Verwendung von Tween in der Waschlösung sowohl auf Glas als auch auf Kunststoff die Hintergrundfluoreszenz reduziert und somit eine Steigerung des Signal/Hintergrundverhältnisses erreicht werden. In den Versuchen zu Ermittlung der Bestimmungsgrenze wurde für das humane Choriongonadotropin (hCG-i) eine Konzentration von 10 mU/ml, für dessen beta-Untereinheit (hCG-beta) eine Konzentration von 3,6 mU/ml und für das luteinisierende Hormon (LH) eine Konzentration von 10 mU/ml bestimmt. Den ermittelten Wert im Serum für das hCG-i entspricht dem von der Welt-Anti-Dopin-Agentur (WADA) geforderten Wert in Urin von 5 mU/ml. Neben der Ermittlung von Bestimmungsgrenzen wurden diese hinsichtlich auftretender Matrixeffekte in Serum und Blut gemessen. Wie aus den Versuchen zur Ermittlung von Kreuzreaktivitäten auf dem Mikroarray zu entnehmen ist, lassen sich das LH, das hCG-i und hCG-β ebenfalls in Serum und Blut messen. Die Durchführung einer Performance-Analyse über einem Einfach-Blind-Ansatz mit 130 Serum Proben, wurde ebenfalls über dieses System realisiert. Die ausgewerteten Proben wurden anschließend über eine Grenzwertoptimierungskurve analysiert und die diagnostische Spezifität ermittelt. Für die Messungen des LH konnte eine Sensitivität und Spezifität von 100% erreicht werden. Demnach wurden alle negativen und positiven Proben eindeutig interpretiert. Für das hCG-β konnte ebenfalls eine Spezifität von 100% und eine Sensitivität von 97% erreicht werden. Die hCG-i Proben wurden mit einer Spezifität von 100% und eine Sensitivität von 97,5% gemessen. Um den Nachweis zu erbringen, dass dieser Versuchsaufbau über mehrere Wochen stabile Signale bei Vermessen von identischen Proben liefert, wurde ein über zwölf Wochen angesetzter Stabilitätstest für alle Parameter erfolgreich in Serum und Blut durchgeführt. Zusammenfassend konnte in dieser Arbeit erfolgreich eine Mehrkomponentenanalyse als Multiplex Ansatz auf einem Mikroarray entwickelt werden. Die Durchführung der Performance-Analyse und des Stabilitätstests zeigen bereits die mögliche Einsatzfähigkeit dieses Tests im Kontext einer Dopinganalyse.
Das Influenzavirus infiziert Säugetiere und Vögel. Der erste Schritt im Infektionszyklus ist die Anbindung des Viruses über sein Oberflächenprotein Hämagglutinin (HA) an Zuckerstrukturen auf Epithelzellen des respiratorischen Traktes im Wirtsorganismus. Aus den drei komplementaritätsbestimmenden Regionen (complementarity determining regions, CDRs) der schweren Kette eines monoklonalen Hämagglutinin-bindenden Antikörpers wurden drei lineare Peptide abgeleitet. Die Bindungseigenschaften der drei Peptide wurden experimentell mittels Oberflächenplasmonenresonanzspektroskopie untersucht. Es zeigte sich, dass in Übereinstimmung mit begleitenden Molekulardynamik-Simulationen zwei der drei Peptide (PeB und PeC) analog zur Bindefähigkeit des Antikörpers in der Lage sind, Influenzaviren vom Stamm X31 (H3N2 A/Aichi/2/1968) zu binden. Die Interaktion des Peptids PeB, welches potentiell mit der konservierten Rezeptorbindestelle im HA interagiert, wurde anschließend näher charakterisiert. Die Detektion der Influenzaviren war unter geeigneten Immobilisationsbedingungen im diagnostisch relevanten Bereich möglich. Die Spezifität der PeB-Virus-Bindung wurde mittels geeigneter Kontrollen auf der Seite des Analyten und des Liganden nachgewiesen. Des Weiteren war das Peptid PeB in der Lage die Bindung von X31-Viren an Mimetika seines natürlichen Rezeptors zu inhibieren, was die spezifische Interaktion mit der Rezeptorbindungsstelle im Hämagglutinin belegt. Anschließend wurde die Primärsequenz von PeB durch eine vollständige Substitutionsanalyse im Microarray-Format hinsichtlich der Struktur-Aktivitäts-Beziehungen charakterisiert. Dies führte außerdem zu verbesserten Peptidvarianten mit erhöhter Affinität und breiterer Spezifität gegen aktuelle Influenzastämme verschiedener Serotypen (z.B. H1N1/2009, H5N1/2004, H7N1/2013). Schließlich konnte durch Verwendung einer in der Primärsequenz angepassten höher affinen Peptidvariante die Influenzainfektion in vitro inhibiert werden. Damit stellen die vom ursprünglichen Peptid PeB abgeleiteten Varianten Rezeptormoleküle in biosensorischen Testsystemen sowie potentielle Wirkstoffe dar.
Ziel der Arbeit war die Entwicklung von farbstoffmarkierten Polymeren, die einen temperaturgetriebenen Knäuel-Kollaps-Phasenübergang in wässriger Lösung ("thermo-responsive Polymere") zeigen und diesen in ein optisches Signal übersetzen können. Solche Polymere unterliegen innerhalb eines kleinen Temperaturintervalls einer massiven Änderung ihres Verhaltens, z B. ihrer Konformation und ihres Quellungsgrads. Diese Änderungen sind mit einem Wechsel der Löseeigenschaften von hydrophil zu hydrophob verbunden. Als Matrixpolymere wurden Poly-N-isopropylacrylamid (polyNIPAm), Poly(oligoethylen-glykolacrylat) (polyOEGA) und Poly(oligoethylenglykolmethacrylat) (polyOEGMA) ein-gesetzt, in die geeignete Farbstoffen durch Copolymerisation eingebaut wurden. Als besonders geeignet, um den Phasenübergang in ein optisches Signal zu übersetzen, erwiesen sich hierfür kompakte, solvatochrome Cumarin- und Naphthalimidderivate. Diese beeinträchtigten weder das Polymerisationsverhalten noch den Phasenübergang, reagierten aber sowohl bezüglich Farbe als auch Fluoreszenz stark auf die Polarität des Lösemittels. Weiterhin wurden Systeme entwickelt, die mittels Energietransfer (FRET) ein an den Phasenübergang gekoppeltes optisches Signal erzeugen. Hierbei wurde ein Cumarin als Donor- und ein Polythiophen als Akzeptorfarbstoff eingesetzt. Es zeigte sich, dass trotz scheinbarer Ähnlichkeit bestimmte Polymere ausgeprägt auf einen Temperaturstimulus mit Änderung ihrer spektralen Eigenschaften reagieren, andere aber nicht. Hierfür wurden die molekularen Ursachen untersucht. Als wahrscheinliche Gründe für das Ausbleiben einer spektralen Änderung in Oligo(ethylenglykol)-basierten Polymeren sind zum einen die fehlende Dehydratationseffektivität infolge des Fehlens eines selbstgenügenden Wasserstoffbrückenbindungsmotivs zu nennen und zum anderen die sterische Abschirmung der Farbstoffe durch die Oligo(ethylenglykol)-Seitenketten. Als Prinzipbeweis für die Nützlichkeit solcher Systeme für die Bioanalytik wurde ein System entwickelt, dass die Löslichkeitseigenschaft eines thermoresponsiven Polymers durch Antikörper-Antigen-Reaktion änderte. Die Bindung selbst kleiner Mengen eines Antikörpers ließ sich so direkt optisch auslesen und war bereits mit dem bloßen Auge zu erkennen.
Epitop-Kartierung von PBP2A und Identifizierung MRSA-spezifischer immunodominanter Peptidsequenzen
(2014)
Im Rahmen der Dissertation wird die Anwendung und Wirkung von Kernelementen des New Public Management (NPM) am Beispiel der Bürgerdienste der sechs europäischen Hauptstädte Berlin, Brüssel, Kopenhagen, Madrid, Prag und Warschau analysiert. Hierbei steht der Vergleich von Hauptstädten der MOE-Staaten mit Hauptstädten alter EU-Mitgliedsstaaten im Vordergrund. Es wird die folgende Forschungshypothese untersucht: Die Verwaltungen in den Hauptstädten der östlichen Mitgliedsstaaten der EU haben in Folge der grundsätzlichen gesellschaftlichen und politischen Umbrüche in den 1990er Jahren bedeutend mehr Kernelemente des NPM beim Neuaufbau ihrer öffentlichen Verwaltungen eingeführt. Durch den folgerichtigen Aufbau kundenorientierter und moderner Verwaltungen sowie der strikten Anwendung der Kernelemente des New Public Management arbeiten die Bürgerdienste in den Hauptstädten östlicher EU-Mitgliedsstaaten effizienter und wirkungsvoller als vergleichbare Bürgerdienste in den Hauptstädten westlicher EU-Mitgliedsstaaten. Zur Überprüfung der Forschungshypothese werden die Vergleichsstädte zunächst den entsprechenden Rechts- und Verwaltungstraditionen (kontinentaleuropäisch deutsch, napoleonisch und skandinavisch) zugeordnet und bezüglich ihrer Ausgangslage zum Aufbau einer modernen Verwaltung (Westeuropäische Verwaltung, Wiedervereinigungsverwaltung und Transformations-verwaltung) kategorisiert. Im Anschluss werden die institutionellen Voraussetzungen hinterfragt, was die deskriptive Darstellung der Stadt- und Verwaltungsgeschichte sowie die Untersuchung von organisatorischen Strukturen der Bürgerdienste, die Anwendung der NPM-Instrumente als auch die Innen- und Außenperspektive des NPM umfasst. Es wird festgestellt, ob und in welcher Form die Bürgerdienste der Vergleichsstädte die Kernelemente des NPM anwenden. Im Anschluss werden die Vergleichsstädte bezüglich der Anwendung der Kernelemente miteinander verglichen, wobei der Fokus auf dem persönlichen Vertriebsweg und der Kundenorientierung liegt. Der folgende Teil der Dissertation befasst sich mit dem Output der Bürgerdienste, der auf operative Resultate untersucht und verglichen wird. Hierbei stellt sich insbesondere die Frage nach den Leistungsmengen und der Produktivität des Outputs. Es werden aber auch die Ergebnisse von Verwaltungsprozessen untersucht, insbesondere in Bezug auf die Kundenorientierung. Hierfür wird ein Effizienzvergleich der Bürgerdienste in den Vergleichsstädten anhand einer relativen Effizienzmessung und der Free Disposal Hull (FDH)-Methode nach Bouckaert durchgeführt. Es ist eine Konzentration auf populäre Dienstleistungen aus dem Portfolio der Bürgerdienste notwendig. Daher werden die vergleichbaren Dienstleistungen Melde-, Personalausweis-, Führerschein- und Reisepass-angelegenheiten unter Einbeziehung des Vollzeitäquivalents zur Berechnung der Effizienz der Bürgerdienste herangezogen. Hierfür werden Daten aus den Jahren 2009 bis 2011 genutzt, die teilweise aus verwaltungsinternen Datenbanken stammen. Anschließend wird der Versuch unternommen, den Outcome in die Effizienzanalyse der Bürgerdienste einfließen zu lassen. In diesem Zusammenhang wird die Anwendbarkeit von verschiedenen erweiterten Best-Practice-Verfahren und auch eine Erweiterung der relativen Effizienzmessung und der FDH-Methode geprüft. Als Gesamtfazit der Dissertation kann festgehalten werden, dass die Bürgerdienste in den untersuchten Hauptstädten der MOE-Staaten nicht mehr Kernelemente des NPM anwenden, als die Hauptstädte der westlichen Mitgliedsstaaten der EU. Im Gegenteil wendet Prag deutlich weniger NPM-Instrumente als andere Vergleichsstädte an, wohingegen Warschau zwar viele NPM-Instrumente anwendet, jedoch immer von einer westeuropäischen Vergleichsstadt übertroffen wird. Auch die Hypothese, dass die Bürgerdienste in den Hauptstädten der MOE-Staaten effizienter arbeiten als vergleichbare Bürgerdienste in den Hauptstädten westlicher EU-Mitgliedsstaaten wurde durch die Dissertation entkräftet. Das Gegenteil ist der Fall, da Prag und Warschau im Rahmen des Effizienzvergleichs lediglich durchschnittliche oder schlechte Performances aufweisen. Die aufgestellte Hypothese ist durch die Forschungsergebnisse widerlegt, lediglich das gute Abschneiden der Vergleichsstadt Warschau bei der Anwendungsanalyse kann einen Teil der These im gewissen Umfang bestätigen.
In der Berufsgruppe der Lehrerinnen und Lehrer besteht eine hohe Prävalenz psychischer und psychosomatischer Erkrankungen. Aus- und Weiterbildungsangebote zur Vermittlung lehrerspezifischer sozialer Kompetenzen spielen eine wichtige Rolle bei der Förderung der Lehrergesundheit. In der vorliegenden Studie wurde das „Lehrer/innen-Coaching nach dem Freiburger Modell“ evaluiert, welches die Kompetenz von Lehrkräften stärken soll, innerhalb der Schule und insbesondere im Unterricht, schwierige interpersonelle Situationen aktiv und konstruktiv zu gestalten. Damit sollen stressbedingte gesundheitliche Belastungen abgebaut und dem Entstehen gravierender psychischer Störungen vorgebeugt werden. In der vorliegenden Arbeit werden zwei modifizierte Versionen dieses Programms erstmalig im Rahmen einer landesweiten Feldstudie untersucht. Die zentralen Evaluationsfragestellungen beziehen sich auf die Effektivität der Intervention als Gesundheitsförderungsmaßnahme (Akzeptanz, Wirksamkeit, Wirksamkeitsvergleich der beiden Interventionsformen im landesweiten Einsatz). Daneben strebt die Studie einen Vergleich mit den Ergebnissen einer Vorgängerstudie sowie die Generierung weiterer Erkenntnisse zum Zusammenhang zwischen Aspekten der sozialen Kompetenz von Lehrkräften und ihrer psychischen Gesundheit an. An der Maßnahme konnten alle baden-württembergischen Lehrerinnen und Lehrer mit einer Berufserfahrung von mindestens 10 Jahren teilnehmen. Für die Untersuchung der Wirksamkeit der Maßnahme und des Wirksamkeitsvergleichs der beiden unterschiedlichen Formen liegt ein quasiexperimentelles Design mit insgesamt zwei Messzeitpunkten vor. In die Auswertung zur Wirksamkeit der Intervention konnten die Daten von den 314 Teilnehmern einbezogen werden. Die Messinstrumente, die in der vorliegenden Studie zur Anwendung kamen, waren der General Health Questionnaire (GHQ-12), das Maslach Burnout Inventory (MBI-D) und die ins Deutsche übersetzte Jefferson Scale of Empathy (JSE) in der an Lehrer adaptierten Form. Die Evaluationsergebnisse zeigen, dass die Teilnahme am „Lehrer/innen-Coaching nach dem Freiburger Modell” mit einer signifikanten Verbesserung der gesundheitsbezogenen abhängigen Variablen einhergeht. Besonders hervorzuheben ist die ausgeprägte Verbesserung der mittels GHQ-12 erfassten psychischen Gesundheit. Das Ergebnis des Prä-Post-Vergleichs der Gesundheitswerte beider Interventionsgruppen bestätigte sich auch im Vergleich zu einer Null-Interventionsgruppe: Entsprechend der Hypothese gab es bei den Teilnehmern eine signifikant stärkere Verbesserung der psychischen Gesundheit als bei den Nicht-Teilnehmern (Null-Interventionsgruppe). Die beiden Interventionsmodi „Kompaktform” und „Kurzform” erwiesen sich im Hinblick auf die Verbesserung der Lehrergesundheit als gleichermaßen wirksam. Zudem zeigen die Ergebnisse der Teilnehmerbefragung, dass die Maßnahme Anklang bei der Zielgruppe fand. Die Akzeptanz durch die Zielgruppe ist für die Wirksamkeit einer auf Freiwilligkeit basierenden verhaltenspräventiven Maßnahme naturgemäß eine essenzielle Voraussetzung. Bei der psychischen Gesundheit der Lehrer bestehen – wie aus weiteren Befunden der Studie ersichtlich – bedeutsame Zusammenhänge zu einer intakten zwischenmenschlichen Beziehung mit den Schülern, einer gelungenen, durch gegenseitige Unterstützung gekennzeichneten Interaktion im Kollegium und einem entsprechend unterstützenden Führungsverhalten der Schulleitung. Dies macht deutlich, welches besondere Gewicht einer gelingenden Beziehungsgestaltung an Schulen und im Unterricht beizumessen ist. Bezüglich der Vorgehensweise in der vorliegenden Untersuchung werden einige methodische Limitationen hinsichtlich des Designs diskutiert. Ergänzend wird im Ausblick der Evaluationsstudie darauf hingewiesen, wie sich durch die Verknüpfung des vorliegenden Programms mit weiteren, auf den Ebenen Verhalten, Verhältnisse und Führung ansetzenden gesundheitspräventiven Maßnahmen, zukünftig die Stärkung der psychischen Gesundheit von Lehrkräften weiter ausbauen ließe.
Nowadays, software systems are getting more and more complex. To tackle this challenge most diverse techniques, such as design patterns, service oriented architectures (SOA), software development processes, and model-driven engineering (MDE), are used to improve productivity, while time to market and quality of the products stay stable. Multiple of these techniques are used in parallel to profit from their benefits. While the use of sophisticated software development processes is standard, today, MDE is just adopted in practice. However, research has shown that the application of MDE is not always successful. It is not fully understood when advantages of MDE can be used and to what degree MDE can also be disadvantageous for productivity. Further, when combining different techniques that aim to affect the same factor (e.g. productivity) the question arises whether these techniques really complement each other or, in contrast, compensate their effects. Due to that, there is the concrete question how MDE and other techniques, such as software development process, are interrelated. Both aspects (advantages and disadvantages for productivity as well as the interrelation to other techniques) need to be understood to identify risks relating to the productivity impact of MDE. Before studying MDE's impact on productivity, it is necessary to investigate the range of validity that can be reached for the results. This includes two questions. First, there is the question whether MDE's impact on productivity is similar for all approaches of adopting MDE in practice. Second, there is the question whether MDE's impact on productivity for an approach of using MDE in practice remains stable over time. The answers for both questions are crucial for handling risks of MDE, but also for the design of future studies on MDE success. This thesis addresses these questions with the goal to support adoption of MDE in future. To enable a differentiated discussion about MDE, the term MDE setting'' is introduced. MDE setting refers to the applied technical setting, i.e. the employed manual and automated activities, artifacts, languages, and tools. An MDE setting's possible impact on productivity is studied with a focus on changeability and the interrelation to software development processes. This is done by introducing a taxonomy of changeability concerns that might be affected by an MDE setting. Further, three MDE traits are identified and it is studied for which manifestations of these MDE traits software development processes are impacted. To enable the assessment and evaluation of an MDE setting's impacts, the Software Manufacture Model language is introduced. This is a process modeling language that allows to reason about how relations between (modeling) artifacts (e.g. models or code files) change during application of manual or automated development activities. On that basis, risk analysis techniques are provided. These techniques allow identifying changeability risks and assessing the manifestations of the MDE traits (and with it an MDE setting's impact on software development processes). To address the range of validity, MDE settings from practice and their evolution histories were capture in context of this thesis. First, this data is used to show that MDE settings cover the whole spectrum concerning their impact on changeability or interrelation to software development processes. Neither it is seldom that MDE settings are neutral for processes nor is it seldom that MDE settings have impact on processes. Similarly, the impact on changeability differs relevantly. Second, a taxonomy of evolution of MDE settings is introduced. In that context it is discussed to what extent different types of changes on an MDE setting can influence this MDE setting's impact on changeability and the interrelation to processes. The category of structural evolution, which can change these characteristics of an MDE setting, is identified. The captured MDE settings from practice are used to show that structural evolution exists and is common. In addition, some examples of structural evolution steps are collected that actually led to a change in the characteristics of the respective MDE settings. Two implications are: First, the assessed diversity of MDE settings evaluates the need for the analysis techniques that shall be presented in this thesis. Second, evolution is one explanation for the diversity of MDE settings in practice. To summarize, this thesis studies the nature and evolution of MDE settings in practice. As a result support for the adoption of MDE settings is provided in form of techniques for the identification of risks relating to productivity impacts.
Reading is a complex cognitive task based on the analyses of visual stimuli. Due to the physiology of the eye, only a small number of letters around the fixation position can be extracted with high visual acuity, while the visibility of words and letters outside this so-called foveal region quickly drops with increasing eccentricity. As a consequence, saccadic eye movements are needed to repeatedly shift the fovea to new words for visual word identification during reading. Moreover, even within a foveated word fixation positions near the word center are superior to other fixation positions for efficient word recognition (O’Regan, 1981; Brysbaert, Vitu, and Schroyens, 1996). Thus, most reading theories assume that readers aim specifically at word centers during reading (for a review see Reichle, Rayner, & Pollatsek, 2003). However, saccades’ landing positions within words during reading are in fact systematically modulated by the distance of the launch site from the word center (McConkie, Kerr, Reddix, & Zola, 1988). In general, it is largely unknown how readers identify the center of upcoming target words and there is no computational model of the sensorimotor translation of the decision for a target word into spatial word center coordinates. Here we present a series of three studies which aim at advancing the current knowledge about the computation of saccade target coordinates during saccade planning in reading. Based on a large corpus analyses, we firstly identified word skipping as a further factor beyond the launch-site distance with a likewise systematic and surprisingly large effect on within-word landing positions. Most importantly, we found that the end points of saccades after skipped word are shifted two and more letters to the left as compared to one-step saccades (i.e., from word N to word N+1) with equal launch-site distances. Then we present evidence from a single saccade experiment suggesting that the word-skipping effect results from highly automatic low-level perceptual processes, which are essentially based on the localization of blank spaces between words. Finally, in the third part, we present a Bayesian model of the computation of the word center from primary sensory measurements of inter-word spaces. We demonstrate that the model simultaneously accounts for launch-site and saccade-type contingent modulations of within-word landing positions in reading. Our results show that the spatial saccade target during reading is the result of complex estimations of the word center based on incomplete sensory information, which also leads to specific systematic deviations of saccades’ landing positions from the word center. Our results have important implications for current reading models and experimental reading research.
Cette recherche a pour objet l'articulation entre les dimensions anthropologiques et sociologiques de l'anthropologie philosophique de Helmuth Plessner (1892-1985). Elles procèdent selon trois axes. Je m'efforce (1) d'offrir une synthèse de l'anthropologie philosophique plessnerienne afin (2) de reconstituer les conditions de possibilité du social au stade humain de l'organique. Le troisième axe (3) correspond, enfin, à l'analyse des limites structurelles du social à partir de ses deux dimensions constitutives : l'individuel (limites ontogénétiques, comportementales et inter-personnelles) et le collectif (limites culturelles, intra- et inter-culturelles).
Magnetite is an iron oxide, which is ubiquitous in rocks and is usually deposited as small nanoparticulate matter among other rock material. It differs from most other iron oxides because it contains divalent and trivalent iron. Consequently, it has a special crystal structure and unique magnetic properties. These properties are used for paleoclimatic reconstructions where naturally occurring magnetite helps understanding former geological ages. Further on, magnetic properties are used in bio- and nanotechnological applications –synthetic magnetite serves as a contrast agent in MRI, is exploited in biosensing, hyperthermia or is used in storage media.
Magnetic properties are strongly size-dependent and achieving size control under preferably mild synthesis conditions is of interest in order to obtain particles with required properties. By using a custom-made setup, it was possible to synthesize stable single domain magnetite nanoparticles with the co-precipitation method. Furthermore, it was shown that magnetite formation is temperature-dependent, resulting in larger particles at higher temperatures. However, mechanistic approaches about the details are incomplete.
Formation of magnetite from solution was shown to occur from nanoparticulate matter rather than solvated ions. The theoretical framework of such processes has only started to be described, partly due to the lack of kinetic or thermodynamic data. Synthesis of magnetite nanoparticles at different temperatures was performed and the Arrhenius plot was used determine an activation energy for crystal growth of 28.4 kJ mol-1, which led to the conclusion that nanoparticle diffusion is the rate-determining step.
Furthermore, a study of the alteration of magnetite particles of different sizes as a function of their storage conditions is presented. The magnetic properties depend not only on particle size but also depend on the structure of the oxide, because magnetite oxidizes to maghemite under environmental conditions. The dynamics of this process have not been well described. Smaller nanoparticles are shown to oxidize more rapidly than larger ones and the lower the storage temperature, the lower the measured oxidation. In addition, the magnetic properties of the altered particles are not decreased dramatically, thus suggesting that this alteration will not impact the use of such nanoparticles as medical carriers.
Finally, the effect of biological additives on magnetite formation was investigated. Magnetotactic bacteria¬¬ are able to synthesize and align magnetite nanoparticles of well-defined size and morphology due to the involvement of special proteins with specific binding properties. Based on this model of morphology control, phage display experiments were performed to determine peptide sequences that preferably bind to (111)-magnetite faces. The aim was to control the shape of magnetite nanoparticles during the formation. Magnetotactic bacteria are also able to control the intracellular redox potential with proteins called magnetochromes. MamP is such a protein and its oxidizing nature was studied in vitro via biomimetic magnetite formation experiments based on ferrous ions. Magnetite and further trivalent oxides were found.
This work helps understanding basic mechanisms of magnetite formation and gives insight into non-classical crystal growth. In addition, it is shown that alteration of magnetite nanoparticles is mainly based on oxidation to maghemite and does not significantly influence the magnetic properties. Finally, biomimetic experiments help understanding the role of MamP within the bacteria and furthermore, a first step was performed to achieve morphology control in magnetite formation via co-precipitation.
Diese Arbeit untersucht, was passiert, wenn in Non-Profit-Organisation (NPO) der Anspruch des Bürgerschaftlichen Engagements auf Praktiken des Freiwilligenmanagements trifft. Ausgangspunkt dieser Fragestellung ist eine doppelte Diagnose: Zum einen setzen NPOs aufgrund mehrerer Faktoren - u.a. Ressourcenknappheit, Wettbewerb und Nachahmungseffekten – vermehrt auf Freiwilligenmanagement. Mit dieser von der BWL inspirierten, aber für NPO entwickelten Personalführungsmethode wollen sie mehr und bessere Freiwillige gewinnen und deren Einsatz effizienter strukturieren. Zum anderen haben sich gleichzeitig viele NPO dem Ziel des bürgerschaftlichen Engagements verschrieben. Damit reagieren sie auf den aus Politik und Wissenschaft zu vernehmenden Anspruch, die Zivilgesellschaft möge die knappen Kassen der öffentlichen Hand kompensieren und das wachsende Partizipationsbedürfnis weiter Teile der Bevölkerung durch eine neue Kultur der Teilhabe der Bürgerinnen und Bürger befriedigen. Bei näherer Betrachtung zeigt sich jedoch: Während Freiwilligenmanagement einer ökonomischen Handlungslogik folgt, ist bürgerschaftliches Engagement Ausdruck einer Handlungslogik der Zivilgesellschaft. Beide sind unter gegenwärtigen Bedingungen weder theoretisch noch praktisch miteinander vereinbar. Um beide Entwicklungen miteinander zu versöhnen, muss Freiwilligenmanagement unter dem Banner des Bürgerschaftlichen neu gedacht werden. Dieses Argument unterfüttert die Arbeit sowohl theoretisch und empirisch. Der Theorieteil gliedert sich in drei Teile. Zunächst wird der Begriff der NPO näher eingegrenzt. Dazu wird die bestehende Literatur zum Dritten Sektor und Non-Profit-Organisationen zu einem operationalisierbaren Begriff von NPO kondensiert. Daran anschließend werden aktuelle Trends im Feld der NPO identifiziert, die zeigen, dass NPO tatsächlich oft von widerstreitenden Handlungslogiken gekennzeichnet sind, darunter eine ökonomische und eine bürgerschaftliche. Die beiden folgenden Kapitel untersuchen dann jeweils eine der beiden Logiken. Zunächst wird das Leitbild des bürgerschaftlichen Engagements als Ausdruck einer zivilgesellschaftlichen Handlungslogik näher definiert. Dabei zeigt sich, dass dieser Begriff oft sehr unscharf verwendet wird. Daher greift die Arbeit auf die politiktheoretische Diskussion um Zivil- und Bürgergesellschaft auf und schmiedet daraus eine qualifizierte Definition von bürgerschaftlichem Engagement, die sich maßgeblich am Ideal von gesellschaftlich-politischer Partizipation und bürgerschaftlicher Kompetenz orientiert. Dem wird im dritten und letzten Kapitel des Theorieteils die ökonomische Handlungslogik in Form der Theorie des Freiwilligenmanagements gegenübergestellt. Bei der Darstellung zeigt sich schnell, dass dessen Grundprinzipien – anders als oft vorgebracht – mit den qualifizierten Idealen von Partizipation und Konkurrenz im Konflikt stehen. In der empirischen Analyse wird dann in den 8 Interviews den Widersprüchen zwischen bürgerschaftlichem Engagement und Freiwilligenmanagement in der Praxis nachgegangen. Die Ergebnisse dieser Untersuchung lassen sich in 5 Punkten zusammenfassen: 1. Freiwilligenmanagement orientiert sich erstens im wesentlichen an einer Zahl: Dem Zugewinn oder Verlust von freiwilliger Arbeit. 2. Freiwilligenmanagement installiert ein umfassendes System der Selektion von „passenden“ Freiwilligen. 3. Positiv hervorzuheben ist die institutionalisierte Ansprechbarkeit, die im Rahmen von Freiwilligenmanagement in NPO Einzug erhält. 4. Freiwilligenmanagement ist eng mit dem Anspruch verbunden, die Arbeit der Freiwilligen zu kontrollieren. Der Eigensinn des Engagements, die Notwendigkeit von Spielräumen, die Möglichkeit des Ausprobierens oder der Anspruch der Freiwilligen, an Entscheidungen zu partizipieren bzw. gar selbstorganisiert und -verantwortlich zu handeln, rückt dabei in den Hintergrund. 5. In den Interviews wird eine starke Ökonomisierung des Engagements sichtbar. Freiwillige werden als Ressource betrachtet, ihr Engagement als „Zeitspende“ statistisch erfasst, ihre (Dienst-)Leistung monetär bewertet. Im Zuge dessen erhält auch der Managerialism verstärkt Einfluss auf die Arbeit in NPO und begründet ein stark hierarchisches Verhältnis: Während die Freiwilligenmangerin aktiv handelt, wird die freiwillig Engagierte zum Objekt von Management-Techniken. Dass dies dem Anspruch der Partizipation entgegenläuft, ergibt sich dabei von selbst. Angesichts dieser Diagnose, dass real-existierendes Freiwilligenmanagement nicht mit dem Ideal des bürgerschaftlichen Engagement im engeren Sinne zusammenpasst, formuliert das Fazit Vorschläge für ein bürgerschaftlich orientiertes, engagement-sensibles Freiwilligenmanagement.
During this work I built a four wave mixing setup for the time-resolved femtosecond spectroscopy of Raman-active lattice modes. This setup enables to study the selective excitation of phonon polaritons. These quasi-particles arise from the coupling of electro-magnetic waves and transverse optical lattice modes, the so-called phonons. The phonon polaritons were investigated in the optically non-linear, ferroelectric crystals LiNbO₃ and LiTaO₃.
The direct observation of the frequency shift of the scattered narrow bandwidth probe pulses proofs the role of the Raman interaction during the probe and excitation process of phonon polaritons. I compare this experimental method with the measurement where ultra-short laser pulses are used. The frequency shift remains obscured by the relative broad bandwidth of these laser pulses. In an experiment with narrow bandwidth probe pulses, the Stokes and anti-Stokes intensities are spectrally separated. They are assigned to the corresponding counter-propagating wavepackets of phonon polaritons. Thus, the dynamics of these wavepackets was separately studied. Based on these findings, I develop the mathematical description of the so-called homodyne detection of light for the case of light scattering from counter propagating phonon polaritons.
Further, I modified the broad bandwidth of the ultra-short pump pulses using bandpass filters to generate two pump pulses with non-overlapping spectra. This enables the frequency-selective excitation of polariton modes in the sample, which allows me to observe even very weak polariton modes in LiNbO₃ or LiTaO₃ that belong to the higher branches of the dispersion relation of phonon polaritons. The experimentally determined dispersion relation of the phonon polaritons could therefore be extended and compared to theoretical models. In addition, I determined the frequency-dependent damping of phonon polaritons.
Knowing the rates and mechanisms of geomorphic process that shape the Earth’s surface is crucial to understand landscape evolution. Modern methods for estimating denudation rates enable us to quantitatively express and compare processes of landscape downwearing that can be traced through time and space—from the seemingly intact, though intensely shattered, phantom blocks of the catastrophically fragmented basal facies of giant rockslides up to denudational noise in orogen-wide data sets averaging over several millennia. This great variety of spatiotemporal scales of denudation rates is both boon and bane of geomorphic process rates. Indeed, processes of landscape downwearing can be traced far back in time, helping us to understand the Earth’s evolution. Yet, this benefit may turn into a drawback due to scaling issues if these rates are to be compared across different observation timescales.
This thesis investigates the mechanisms, patterns and rates of landscape downwearing across the Himalaya-Tibet orogen.
Accounting for the spatiotemporal variability of denudation processes, this thesis addresses landscape downwearing on three distinctly different spatial scales, starting off at the local scale of individual hillslopes where considerable amounts of debris are generated from rock instantaneously: Rocksliding in active mountains is a major impetus of landscape downwearing. Study I provides a systematic overview of the internal sedimentology of giant rockslide deposits and thus meets the challenge of distinguishing them from macroscopically and microscopically similar glacial deposits, tectonic fault-zone breccias, and impact breccias. This distinction is important to avoid erroneous or misleading deduction of paleoclimatic or tectonic implications. -> Grain size analysis shows that rockslide-derived micro-breccia closely resemble those from meteorite impact or tectonic faults. -> Frictionite may occur more frequently that previously assumed. -> Mössbauer-spectroscopy derived results indicate basal rock melting in the absence of water, involving short-term temperatures of >1500°C.
Zooming out, Study II tracks the fate of these sediments, using the example of the upper Indus River, NW India. There we use river sand samples from the Indus and its tributaries to estimate basin-averaged denudation rates along a ~320-km reach across the Tibetan Plateau margin, to answer the question whether incision into the western Tibetan Plateau margin is currently active or not. -> We find an about one-order-of-magnitude upstream decay—from 110 to 10 mm kyr^-1—of cosmogenic Be-10-derived basin-wide denudation rates across the morphological knickpoint that marks the transition from the Transhimalayan ranges to the Tibetan Plateau. This trend is corroborated by independent bulk petrographic and heavy mineral analysis of the same samples. -> From the observation that tributary-derived basin-wide denudation rates do not increase markedly until ~150–200 km downstream of the topographic plateau margin we conclude that incision into the Tibetan Plateau is inactive. -> Comparing our postglacial Be-10-derived denudation rates to long-term (>10^6 yr) estimates from low-temperature thermochronometry, ranging from 100 to 750 mm kyr^-1, points to an order- of-magnitude decay of rates of landscape downwearing towards present. We infer that denudation rates must have been higher in the Quaternary, probably promoted by the interplay of glacial and interglacial stages.
Our investigation of regional denudation patterns in the upper Indus finally is an integral part of Study III that synthesizes denudation of the Himalaya-Tibet orogen. In order to identify general and time-invariant predictors for Be-10-derived denudation rates we analyze tectonic, climatic and topographic metrics from an inventory of 297 drainage basins from various parts of the orogen. Aiming to get insight to the full response distributions of denudation rate to tectonic, climatic and topographic candidate predictors, we apply quantile regression instead of ordinary least squares regression, which has been standard analysis tool in previous studies that looked for denudation rate predictors. -> We use principal component analysis to reduce our set of 26 candidate predictors, ending up with just three out of these: Aridity Index, topographic steepness index, and precipitation of the coldest quarter of the year. -> Topographic steepness index proves to perform best during additive quantile regression. Our consequent prediction of denudation rates on the basin scale involves prediction errors that remain between 5 and 10 mm kyr^-1. -> We conclude that while topographic metrics such as river-channel steepness and slope gradient—being representative on timescales that our cosmogenic Be-10-derived denudation rates integrate over—generally appear to be more suited as predictors than climatic and tectonic metrics based on decadal records.
Galaxies are observational probes to study the Large Scale Structure. Their gravitational motions are tracers of the total matter density and therefore of the Large Scale Structure. Besides, studies of structure formation and galaxy evolution rely on numerical cosmological simulations. Still, only one universe observable from a given position, in time and space, is available for comparisons with simulations. The related cosmic variance affects our ability to interpret the results. Simulations constrained by observational data are a perfect remedy to this problem. Achieving such simulations requires the projects Cosmic flows and CLUES. Cosmic flows builds catalogs of accurate distance measurements to map deviations from the expansion. These measures are mainly obtained with the galaxy luminosity-rotation rate correlation. We present the calibration of that relation in the mid-infrared with observational data from Spitzer Space Telescope. Resulting accurate distance estimates will be included in the third catalog of the project. In the meantime, two catalogs up to 30 and 150 Mpc/h have been released. We report improvements and applications of the CLUES' method on these two catalogs. The technique is based on the constrained realization algorithm. The cosmic displacement field is computed with the Zel'dovich approximation. This latter is then reversed to relocate reconstructed three-dimensional constraints to their precursors' positions in the initial field. The size of the second catalog (8000 galaxies within 150 Mpc/h) highlighted the importance of minimizing the observational biases. By carrying out tests on mock catalogs, built from cosmological simulations, a method to minimize observational bias can be derived. Finally, for the first time, cosmological simulations are constrained solely by peculiar velocities. The process is successful as resulting simulations resemble the Local Universe. The major attractors and voids are simulated at positions approaching observational positions by a few megaparsecs, thus reaching the limit imposed by the linear theory.
Transcription factors (TFs) are ubiquitous gene expression regulators and play essential roles in almost all biological processes. This Ph.D. project is primarily focused on the functional characterisation of MYB112 - a member of the R2R3-MYB TF family from the model plant Arabidopsis thaliana. This gene was selected due to its increased expression during senescence based on previous qRT-PCR expression profiling experiments of 1880 TFs in Arabidopsis leaves at three developmental stages (15 mm leaf, 30 mm leaf and 20% yellowing leaf). MYB112 promoter GUS fusion lines were generated to further investigate the expression pattern of MYB112. Employing transgenic approaches in combination with metabolomics and transcriptomics we demonstrate that MYB112 exerts a major role in regulation of plant flavonoid metabolism. We report enhanced and impaired anthocyanin accumulation in MYB112 overexpressors and MYB112-deficient mutants, respectively. Expression profiling reveals that MYB112 acts as a positive regulator of the transcription factor PAP1 leading to increased anthocyanin biosynthesis, and as a negative regulator of MYB12 and MYB111, which both control flavonol biosynthesis. We also identify MYB112 early responsive genes using a combination of several approaches. These include gene expression profiling (Affymetrix ATH1 micro-arrays and qRT-PCR) and transactivation assays in leaf mesophyll cell protoplasts. We show that MYB112 binds to an 8-bp DNA fragment containing the core sequence (A/T/G)(A/C)CC(A/T)(A/G/T)(A/C)(T/C). By electrophoretic mobility shift assay (EMSA) and chromatin immunoprecipitation coupled to qPCR (ChIP-qPCR) we demonstrate that MYB112 binds in vitro and in vivo to MYB7 and MYB32 promoters revealing them as direct downstream target genes. MYB TFs were previously reported to play an important role in controlling flavonoid biosynthesis in plants. Many factors acting upstream of the anthocyanin biosynthesis pathway show enhanced expression levels during nitrogen limitation, or elevated sucrose content. In addition to the mentioned conditions, other environmental parameters including salinity or high light stress may trigger anthocyanin accumulation. In contrast to several other MYB TFs affecting anthocyanin biosynthesis pathway genes, MYB112 expression is not controlled by nitrogen limitation, or carbon excess, but rather is stimulated by salinity and high light stress. Thus, MYB112 constitutes a previously uncharacterised regulatory factor that modifies anthocyanin accumulation under conditions of abiotic stress.