530 Physik
Refine
Year of publication
Document Type
- Article (957)
- Doctoral Thesis (376)
- Postprint (123)
- Preprint (50)
- Other (49)
- Habilitation Thesis (23)
- Review (11)
- Master's Thesis (10)
- Conference Proceeding (5)
- Monograph/Edited Volume (4)
Keywords
- diffusion (45)
- anomalous diffusion (35)
- gamma rays: general (20)
- synchronization (19)
- organic solar cells (17)
- Synchronisation (16)
- stochastic processes (15)
- cosmic rays (14)
- ISM: supernova remnants (13)
- data analysis (12)
Institute
- Institut für Physik und Astronomie (1474)
- Interdisziplinäres Zentrum für Dynamik komplexer Systeme (52)
- Institut für Chemie (45)
- Extern (44)
- Mathematisch-Naturwissenschaftliche Fakultät (25)
- Institut für Mathematik (22)
- Institut für Geowissenschaften (15)
- Institut für Biochemie und Biologie (7)
- Institut für Umweltwissenschaften und Geographie (7)
- Potsdam Institute for Climate Impact Research (PIK) e. V. (6)
Contents: 1 Introduction 1.1 Tikhanov-Phillips Regularization of Ill-Posed Problems 1.2 A Compact Course to Wavelets 2 A Multilevel Iteration for Tikhonov-Phillips Regularization 2.1 Multilevel Splitting 2.2 The Multilevel Iteration 2.3 Multilevel Approach to Cone Beam Reconstuction 3 The use of approximating operators 3.1 Computing approximating families {Ah}
We have shown that the two-dimensional complex Ginzburg-Landau equation exhibits supertransient chaos in a certain parameter range. Using numerical methods this behavior is found near the transition line separating frozen spiral solutions from turbulence. Supertransient chaos seems to be a common phenomenon in extended spatiotemporal systems. These supertransients are characterized by an average transient lifetime which depends exponentially on the size of the system and are due to an underlying nonattracting chaotic set.
Using a special technique of data analysis, we have found out 34 grand minima of solar activity obtained from a 7,700 years long Δ14C record. The method used rests on a proper filtering of the Δ14C record and the extrapolation of verifiable results for the later history back in time. Additionally, we use a method of nonlinear dynamics, the recurrence rate, to back up the results. Our findings are not contradictory to the record of solar maxima resp. minima by Eddy [5], but constitute a considerable extension. Hence, it has become possible to look closer at the validity of models. This way, we have tested several models for solar activity, esp. the model of Barnes et al. [1]. There are hints for that the grand minima might solely be driven by the 209 year period found in the Δ14C record.
Contents: I. Algorithms 1. Theoretical Backround 2. Numerical Procedures 3. Graph Representation of the Solutions 4. Applications and Example II. Users' Manual 5. About the Program 6. The Course of a Qualitative Analysis 7. The Model Module 8. Input description 9. Output Description 10. Example 11. Graphics
It is shown that the ff effect of mean-field magnetohydrodynamics, which consists in the generation of a mean electromotive force along the mean magnetic field by turbulently fluctuating parts of velocity and magnetic field, is equivalent to the simultaneous generation of both turbulent and mean-field magnetic helicities, the generation rates being equal in magnitude and opposite in sign. In the particular case of statistically stationary and homogeneous fluctuations this implies that the ff effect can increase the energy in the mean magnetic field only under the condition that also magnetic helicity is accumulated there.
We report on bifurcation studies for the incompressible Navier-Stokes equations in two space dimensions with periodic boundary conditions and an external forcing of the Kolmogorov type. Fourier representations of velocity and pressure have been used to approximate the original partial differential equations by a finite-dimensional system of ordinary differential equations, which then has been studied by means of bifurcation-analysis techniques. A special route into chaos observed for increasing Reynolds number or strength of the imposed forcing is described. It includes several steady states, traveling waves, modulated traveling waves, periodic and torus solutions, as well as a period-doubling cascade for a torus solution. Lyapunov exponents and Kaplan-Yorke dimensions have been calculated to characterize the chaotic branch. While studying the dynamics of the system in Fourier space, we also have transformed solutions to real space and examined the relation between the different bifurcations in Fourier space and toplogical changes of the streamline portrait. In particular, the time-dependent solutions, such as, e.g., traveling waves, torus, and chaotic solutions, have been characterized by the associated fluid-particle motion (Lagrangian dynamics).
Projection methods based on wavelet functions combine optimal convergence rates with algorithmic efficiency. The proofs in this paper utilize the approximation properties of wavelets and results from the general theory of regularization methods. Moreover, adaptive strategies can be incorporated still leading to optimal convergence rates for the resulting algorithms. The so-called wavelet-vaguelette decompositions enable the realization of especially fast algorithms for certain operators.
The bifurcation behaviour of the 3D magnetohydrodynamic equations has been studied for external forcings of varying degree of helicity. With increasing strength of the forcing a primary non-magnetic steady state loses stability to a magnetic periodic state if the helicity exceeds a threshold value and to different non-magnetic states otherwise.
The present paper is related to the problem of approximating the exact solution to the magnetohydrodynamic equations (MHD). The behaviour of a viscous, incompressible and resistive fluid is exemined for a long period of time. Contents: 1 The magnetohydrodynamic equations 2 Notations and precise functional setting of the problem 3 Existence, uniqueness and regularity results 4 Statement and Proof of the main theorem 5 The approximate inertial manifold 6 Summary
We demonstrate the occurrence of regimes with singular continuous (fractal) Fourier spectra in autonomous dissipative dynamical systems. The particular example in an ODE system at the accumulation points of bifurcation sequences associated to the creation of complicated homoclinic orbits. Two different machanisms responsible for the appearance of such spectra are proposed. In the first case when the geometry of the attractor is symbolically represented by the Thue-Morse sequence, both the continuous-time process and its descrete Poincaré map have singular power spectra. The other mechanism owes to the logarithmic divergence of the first return times near the saddle point; here the Poincaré map possesses the discrete spectrum, while the continuous-time process displays the singular one. A method is presented for computing the multifractal characteristics of the singular continuous spectra with the help of the usual Fourier analysis technique.
The Voyager 2 Photopolarimeter experiment has yielded the highest resolved data of Saturn's rings, exhibiting a wide variety of features. The B-ring region between 105000 km and 110000 km distance from Saturn has been investigated. It has a high matter density and contains no significance features visible by eye. Analysis with statistical methods has let us to the detection of two significant events. These features are correlated with the inner 3:2 resonances of the F-ring shepherd satellites Pandora and Prometheus, and may be evidence of large ring paricles caught in the corotation resonances.
We report on bifurcation studies for the incompressible magnetohydrodynamic equations in three space dimensions with periodic boundary conditions and a temporally constant external forcing. Fourier reprsentations of velocity, pressure and magnetic field have been used to transform the original partial differential equations into systems of ordinary differential equations (ODE), to which then special numerical methods for the qualitative analysis of systems of ODE have been applied, supplemented by the simulative calculation of solutions for selected initial conditions. In a part of the calculations, in order to reduce the number of modes to be retained, the concept of approximate inertial manifolds has been applied. For varying (incereasing from zero) strength of the imposed forcing, or varying Reynolds number, respectively, time-asymptotic states, notably stable stationary solutions, have been traced. A primary non-magnetic steady state loses, in a Hopf bifurcation, stability to a periodic state with a non-vanishing magnetic field, showing the appearance of a generic dynamo effect. From now on the magnetic field is present for all values of the forcing. The Hopf bifurcation is followed by furhter, symmetry-breaking, bifurcations, leading finally to chaos. We pay particular attention to kinetic and magnetic helicities. The dynamo effect is observed only if the forcing is chosen such that a mean kinetic helicity is generated; otherwise the magnetic field diffuses away, and the time-asymptotic states are non-magnetic, in accordance with traditional kinematic dynamo theory.
It is desirable to reduce the potential threats that result from the variability of nature, such as droughts or heat waves that lead to food shortage, or the other extreme, floods that lead to severe damage. To prevent such catastrophic events, it is necessary to understand, and to be capable of characterising, nature's variability. Typically one aims to describe the underlying dynamics of geophysical records with differential equations. There are, however, situations where this does not support the objectives, or is not feasible, e.g., when little is known about the system, or it is too complex for the model parameters to be identified. In such situations it is beneficial to regard certain influences as random, and describe them with stochastic processes. In this thesis I focus on such a description with linear stochastic processes of the FARIMA type and concentrate on the detection of long-range dependence. Long-range dependent processes show an algebraic (i.e. slow) decay of the autocorrelation function. Detection of the latter is important with respect to, e.g. trend tests and uncertainty analysis. Aiming to provide a reliable and powerful strategy for the detection of long-range dependence, I suggest a way of addressing the problem which is somewhat different from standard approaches. Commonly used methods are based either on investigating the asymptotic behaviour (e.g., log-periodogram regression), or on finding a suitable potentially long-range dependent model (e.g., FARIMA[p,d,q]) and test the fractional difference parameter d for compatibility with zero. Here, I suggest to rephrase the problem as a model selection task, i.e.comparing the most suitable long-range dependent and the most suitable short-range dependent model. Approaching the task this way requires a) a suitable class of long-range and short-range dependent models along with suitable means for parameter estimation and b) a reliable model selection strategy, capable of discriminating also non-nested models. With the flexible FARIMA model class together with the Whittle estimator the first requirement is fulfilled. Standard model selection strategies, e.g., the likelihood-ratio test, is for a comparison of non-nested models frequently not powerful enough. Thus, I suggest to extend this strategy with a simulation based model selection approach suitable for such a direct comparison. The approach follows the procedure of a statistical test, with the likelihood-ratio as the test statistic. Its distribution is obtained via simulations using the two models under consideration. For two simple models and different parameter values, I investigate the reliability of p-value and power estimates obtained from the simulated distributions. The result turned out to be dependent on the model parameters. However, in many cases the estimates allow an adequate model selection to be established. An important feature of this approach is that it immediately reveals the ability or inability to discriminate between the two models under consideration. Two applications, a trend detection problem in temperature records and an uncertainty analysis for flood return level estimation, accentuate the importance of having reliable methods at hand for the detection of long-range dependence. In the case of trend detection, falsely concluding long-range dependence implies an underestimation of a trend and possibly leads to a delay of measures needed to take in order to counteract the trend. Ignoring long-range dependence, although present, leads to an underestimation of confidence intervals and thus to an unjustified belief in safety, as it is the case for the return level uncertainty analysis. A reliable detection of long-range dependence is thus highly relevant in practical applications. Examples related to extreme value analysis are not limited to hydrological applications. The increased uncertainty of return level estimates is a potentially problem for all records from autocorrelated processes, an interesting examples in this respect is the assessment of the maximum strength of wind gusts, which is important for designing wind turbines. The detection of long-range dependence is also a relevant problem in the exploration of financial market volatility. With rephrasing the detection problem as a model selection task and suggesting refined methods for model comparison, this thesis contributes to the discussion on and development of methods for the detection of long-range dependence.
In the modern industrialized countries every year several hundred thousands of people die due to the sudden cardiac death. The individual risk for this sudden cardiac death cannot be defined precisely by common available, non-invasive diagnostic tools like Holter-monitoring, highly amplified ECG and traditional linear analysis of heart rate variability (HRV). Therefore, we apply some rather unconventional methods of nonlinear dynamics to analyse the HRV. Especially, some complexity measures that are basing on symbolic dynamics as well as a new measure, the renormalized entropy, detect some abnormalities in the HRV of several patients who have been classified in the low risk group by traditional methods. A combination of these complexity measures with the parameters in the frequency domain seems to be a promising way to get a more precise definition of the individual risk. These findings have to be validated by a representative number of patients.
We have used techniques of nonlinear dynamics to compare a special model for the reversals of the Earth's magnetic field with the observational data. Although this model is rather simple, there is no essential difference to the data by means of well-known characteristics, such as correlation function and probability distribution. Applying methods of symbolic dynamics we have found that the considered model is not able to describe the dynamical properties of the observed process. These significant differences are expressed by algorithmic complexity and Renyi information.
Two deterministic processes leading to roughening interfaces are considered. It is shown that the dynamics of linear perturbations of turbulent regimes in coupled map lattices is governed by a discrete version of the Kardar-Parisi-Zhang equation. The asymptotic scaling behavior of the perturbation field is investigated in the case of large lattices. Secondly, the dynamics of an order-disorder interface is modelled with a simple two-dimensional coupled map lattice, possesing a turbulent and a laminar state. It is demonstrated, that in some range of parameters the spreading of the turbulent state is accompanied by kinetic roughening of the interface.
Strange nonchaotic attractors typically appear in quasiperiodically driven nonlinear systems. Two methods of their characterization are proposed. The first one is based on the bifurcation analysis of the systems, resulting from periodic approximations of the quasiperiodic forcing. Secondly, we propose th characterize their strangeness by calculating a phase sensitivity exponent, that measures the sensitivity with respect to changes of the phase of the external force. It is shown, that phase sensitivity appears if there is a non-zero probability for positive local Lyapunov exponents to occur.
We have studied bifurcation phenomena for the incompressable Navier-Stokes equations in two space dimensions with periodic boundary conditions. Fourier representations of velocity and pressure have been used to transform the original partial differential equations into systems of ordinary differential equations (ODE), to which then numerical methods for the qualitative analysis of systems of ODE have been applied, supplemented by the simulative calculation of solutions for selected initial conditions. Invariant sets, notably steady states, have been traced for varying Reynolds number or strength of the imposed forcing, respectively. A complete bifurcation sequence leading to chaos is described in detail, including the calculation of the Lyapunov exponents that characterize the resulting chaotic branch in the bifurcation diagram.
Nowadays, colloidal rods can be synthesized in large amounts. The rods are typically cylindrically and their length ranges from several nanometers to a few micrometers. In solution, systems of colloidal rodlike molecules or aggregates can form liquid-crystalline phases with long-range orientational and spatial order. In the present work, we investigate structure formation and fractionation in systems of rodlike colloids with the help of Monte Carlo simulations in the NPT ensemble. Repulsive interactions can successfully be mimicked by the hard rod model, which has been studied extensively in the past. In many cases, attractive interactions like van der Waals or depletion forces cannot be neglected, however. In the first part of this work, the phase behavior of monodisperse attractive rods is characterized for different interaction strengths. Phase diagrams as a function of rod length and pressure are presented. Most systems of synthesized mesoscopic rods have a polydisperse length distribution as a consequence of the longitudinal growth process of the rods. For many technical and research applications, a rather small polydispersity is desired in order to have well defined material properties. The polydispersity can be reduced by a spatial demixing (fractionation) of long and short rods. Fractionation and structure formation is studied in a tridisperse and a polydisperse bulk suspension of rods. We observe that the resulting structures depend distinctly on the interaction strength. The fractionation in the system is strongly enhanced with increasing interaction strength. Suspensions are typically confined in a container. We also examine the influence of adjacent substrates in systems of tridisperse and polydisperse rod suspensions. Three different substrate types are studied in detail: a planar wall, a corrugated substrate, and a substrate with rectangular cavities. We analyze the fluid structure close to the substrate and substrate controlled fractionation. The spatial arrangement of long and short rods in front of the substrate depends sensitively on the substrate structure and the pressure. Rods with a predefined length are segregated at substrates with rectangular cavities.
In this thesis the interplay between hydrodynamic transport and specific adhesion is theoretically investigated. An important biological motivation for this work is the rolling adhesion of white blood cells experimentally investigated in flow chambers. There, specific adhesion is mediated by weak bonds between complementary molecular building blocks which are either located on the cell surface (receptors) or attached to the bottom plate of the flow chamber (ligands). The model system under consideration is a hard sphere covered with receptors moving above a planar ligand-bearing wall. The motion of the sphere is influenced by a simple shear flow, deterministic forces, and Brownian motion. An algorithm is given that allows to numerically simulate this motion as well as the formation and rupture of bonds between receptors and ligands. The presented algorithm spatially resolves receptors and ligands. This opens up the perspective to apply the results also to flow chamber experiments done with patterned substrates based on modern nanotechnological developments. In the first part the influence of flow rate, as well as of the number and geometry of receptors and ligands, on the probability for initial binding is studied. This is done by determining the mean time that elapses until the first encounter between a receptor and a ligand occurs. It turns out that besides the number of receptors, especially the height by which the receptors are elevated above the surface of the sphere plays an important role. These findings are in good agreement with observations of actual biological systems like white blood cells or malaria-infected red blood cells. Then, the influence of bonds which have formed between receptors and ligands, but easily rupture in response to force, on the motion of the sphere is studied. It is demonstrated that different states of motion-for example rolling-can be distinguished. The appearance of these states depending on important model parameters is then systematically investigated. Furthermore, it is shown by which bond property the ability of cells to stably roll in a large range of applied flow rates is increased. Finally, the model is applied to another biological process, the transport of spherical cargo particles by molecular motors. In analogy to the so far described systems molecular motors can be considered as bonds that are able to actively move. In this part of the thesis the mean distance the cargo particles are transported is determined.
The separation of natural and anthropogenically caused climatic changes is an important task of contemporary climate research. For this purpose, a detailed knowledge of the natural variability of the climate during warm stages is a necessary prerequisite. Beside model simulations and historical documents, this knowledge is mostly derived from analyses of so-called climatic proxy data like tree rings or sediment as well as ice cores. In order to be able to appropriately interpret such sources of palaeoclimatic information, suitable approaches of statistical modelling as well as methods of time series analysis are necessary, which are applicable to short, noisy, and non-stationary uni- and multivariate data sets. Correlations between different climatic proxy data within one or more climatological archives contain significant information about the climatic change on longer time scales. Based on an appropriate statistical decomposition of such multivariate time series, one may estimate dimensions in terms of the number of significant, linear independent components of the considered data set. In the presented work, a corresponding approach is introduced, critically discussed, and extended with respect to the analysis of palaeoclimatic time series. Temporal variations of the resulting measures allow to derive information about climatic changes. For an example of trace element abundances and grain-size distributions obtained near the Cape Roberts (Eastern Antarctica), it is shown that the variability of the dimensions of the investigated data sets clearly correlates with the Oligocene/Miocene transition about 24 million years before present as well as regional deglaciation events. Grain-size distributions in sediments give information about the predominance of different transportation as well as deposition mechanisms. Finite mixture models may be used to approximate the corresponding distribution functions appropriately. In order to give a complete description of the statistical uncertainty of the parameter estimates in such models, the concept of asymptotic uncertainty distributions is introduced. The relationship with the mutual component overlap as well as with the information missing due to grouping and truncation of the measured data is discussed for a particular geological example. An analysis of a sequence of grain-size distributions obtained in Lake Baikal reveals that there are certain problems accompanying the application of finite mixture models, which cause an extended climatological interpretation of the results to fail. As an appropriate alternative, a linear principal component analysis is used to decompose the data set into suitable fractions whose temporal variability correlates well with the variations of the average solar insolation on millenial to multi-millenial time scales. The abundance of coarse-grained material is obviously related to the annual snow cover, whereas a significant fraction of fine-grained sediments is likely transported from the Taklamakan desert via dust storms in the spring season.
Variationen der stratosphärischen Residualzirkulation und ihr Einfluss auf die Ozonverteilung
(2006)
Die Residualzirkulation entspricht der mittleren Massenzirkulation und beschreibt die im zonalen Mittel stattfindenden meridionalen Transportprozesse. Die Variationen der Residualzirkulation bestimmen gemeinsam mit dem anthropogen verursachten Ozonabbau die jährlichen Schwankungen der Ozongesamtsäule im arktischen Frühling. In der vorliegenden Arbeit wird die Geschwindigkeit des arktischen Astes der Residualzirkulation aus atmosphärischen Daten gewonnen. Zu diesem Zweck wird das diabatische Absinken im Polarwirbel mit Hilfe von Trajektorienrechnungen bestimmt. Die vertikalen Bewegungen der Luftpakete können mit vertikalen Windfeldern oder entsprechend einem neuen Ansatz mit diabatischen Heizraten angetrieben werden. Die Eingabedaten stammen aus dem 45 Jahre langen Reanalyse-Datensatz des "European Centre for Medium Range Weather Forecast" (ECMWF). Außerdem kann für die Jahre ab 1984 die operationelle ECMWF-Analyse verwendet werden. Die Qualität und Robustheit der Heizraten- und Trajektorienrechnungen werden durch Sensitivitätsstudien und Vergleiche mit anderen Modellen untermauert. Anschließend werden umfangreiche Trajektorienensemble statistisch ausgewertet, um ein detailliertes, zeit- und höhenaufgelöstes Bild des diabatischen Absinkens zu ermitteln. In diesem Zusammenhang werden zwei Methoden entwickelt, um das Absinken gemittelt im Polarwirbel oder als Funktion der äquivalenten Breite zu bestimmen. Es wird gezeigt, dass es notwendig ist den Lagrangeschen auf Trajektorienrechnungen basierenden Ansatz zu verfolgen, da die einfachen Eulerschen Mittel Abweichungen zu den Lagrangeschen Vertikalgeschwindigkeiten aufweisen. Das wirbelgemittelte Absinken wird für einzelne Winter mit dem beobachteten Absinken langlebiger Spurengase und anderen Modellstudien verglichen. Der Vergleich zeigt, dass das Absinken basierend auf den vertikalen Windfeldern der ECMWF-Datensätze den Nettoluftmassentransport durch die Residualzirkulation sehr stark überschätzt. Der neue Ansatz basierend auf den Heizraten ergibt hingegen realistische Ergebnisse und wird aus diesem Grund für alle Rechnungen verwendet. Es wird erstmalig eine Klimatologie des diabatischen Absinkens über einen fast fünf Jahrzehnte umfassenden Zeitraum erstellt. Die Klimatologie beinhaltet das vertikal und zeitlich aufgelöste diabatische Absinken gemittelt über den gesamten Polarwirbel und Informationen über die räumliche Struktur des vertikalen Absinkens. Die natürliche Jahr-zu-Jahr Variabilität des diabatischen Absinkens ist sehr stark ausgeprägt. Es wird gezeigt, dass zwischen der ECMWF-Zeitreihe des diabatischen Absinkens und der Zeitreihe aus einem unabhängig analysierten Temperaturdatensatz hohe Korrelationen bestehen. Erstmals wird der Einfluss von Transportprozessen auf die Ozongesamtsäule im arktischen Frühling direkt quantifiziert. Es wird gezeigt, dass die Jahr-zu-Jahr Variabilität der Ozongesamtsäule im arktischen Frühling zu gleichen Anteilen durch die Variabilität der dynamischen Komponente und durch die Variabilität der chemischen Komponente beeinflusst wird. Die gefundenen Variabilitäten von diabatischem Absinken und Ozoneintrag in hohen Breiten werden mit der vertikalen Ausbreitung planetarer Wellen aus der Troposphäre in die Stratosphäre in Beziehung gesetzt.
The primary objective of this work was to develop a laser source for fundamental investigations in the field of laser – materials interactions. In particular it is supposed to facilitate the study of the influence of the temporal energy distribution such as the interaction between adjacent pulses on ablation processes. Therefore, the aim was to design a laser with a highly flexible and easily controllable temporal energy distribution. The laser to meet these demands is an SBS-laser with optional active mode-locking. The nonlinear reflectivity of the SBS-mirror leads to a passive Q-switching and issues ns-pulse bursts with µs spacing. The pulse train parameters such as pulse duration, pulse spacing, pulse energy and number of pulses within a burst can be individually adjusted by tuning the pump parameters and the starting conditions for the laser. Another feature of the SBS-reflection is phase conjugation, which leads to an excellent beam quality thanks to the compensation of phase distortions. Transverse fundamental mode operation and a beam quality better than 1.4 times diffraction limited can be maintained for average output powers of up to 10 W. In addition to the dynamics on a ns-timescale described above, a defined splitting up of each ns-pulse into a train of ps-pulses can be achieved by additional active mode-locking. This twofold temporal focussing of the intensity leads to single pulse energies of up to 2 mJ at pulse durations of approximately 400 ps which corresponds to a pulse peak power of 5 MW. While the pulse duration is of the same order of magnitude as those of other passively Q-switched lasers with simultaneous mode-locking, the pulse energy and pulse peak power exceeds the values of these systems found in the literature by an order of magnitude. To the best of my knowledge the laser presented here is the first implementation of a self-starting mode-locked SBS-laser oscillator. In order to gain a better understanding and control of the transient output of the laser two complementary numerical models were developed. The first is based on laser rate equations which are solved for each laser mode individually while the mode-locking dynamics are calculated from the resultant transient spectrum. The rate equations consider the mean photon densities in the resonator, therefore the propagation of the light inside the resonator is not properly displayed. The second model, in contrast, introduces a spatial resolution of the resonator and hence the propagation inside the resonator can more accurately be considered. Consequently, a mismatch between the loss modulation frequency and the resonator round trip time can be conceived. The model calculates all dynamics in the time domain and therefore the spectral influences such as the Stokes-shift have to be neglected. Both models achieve an excellent reproduction of the ns-dynamics that are generated by the SBS-Q-switch. Separately, each model fails to reproduce all aspects of the ps-dynamics of the SBS-laser in detail. This can be attributed to the complexity of the numerous physical processes involved in this system. But thanks to their complementary nature they provide a very useful tool for investigating the various influences on the dynamics of the mode-locked SBS-laser individually. These aspects can eventually be recomposed to give a complete picture of the mechanisms which govern the output dynamics. Among the aspects under scrutiny were in particular the start resonator quality which determines the starting condition for the SBS-Q-switch, the modulation depth of the AOM and the phonon lifetime as well as the Brillouin-frequency of the SBS-medium. The numerical simulations and the experiments have opened several doors inviting further investigations and promising a potential for further improvement of the experimental results: The results of the simulations in combination with the experimental results which determined the starting conditions for the simulations leave no doubt that the bandwidth generation can primarily be attributed to the SBS-Stokes-shift during the buildup of the Q-switch pulse. For each resonator round trip, bandwidth is generated by shifting a part of the revolving light in frequency. The magnitude of the frequency shift corresponds to the Brillouin-frequency which is a constant of the SBS material and amounts in the case of SF6 to 240 MHz. The modulation of the AOM merely provides an exchange of population between spectrally adjacent modes and therefore diminishes a modulation in the spectrum. By use of a material with a Brillouin-frequency in the GHz range the bandwidth generation can be considerably accelerated thereby shortening the pulse duration. Also, it was demonstrated that yet another nonlinear effect of the SBS can be exploited: If the phonon lifetime is short compared to the resonator round trip time we obtain a modulation in the SBS-reflectivity that supports the modulation of the AOM. The application of an external optical feedback by a conventional mirror turns out to be an alternative to the AOM in synchronizing the longitudinal resonator modes. The interesting feature about this system is that it is ― although highly complex in the physical processes and the temporal output dynamics ― very simple and inexpensive from a technical point of view. No expensive modulators and no control electronics are necessary. Finally, the numerical models constitute a powerful tool for the investigation of emission dynamics of complex laser systems on arbitrary timescales and can also display the spectral evolution of the laser output. In particular it could be demonstrated that differences in the results of the complementary models vanish for systems of lesser complexity.
Uncertainties are pervasive in the Earth System modelling. This is not just due to a lack of knowledge about physical processes but has its seeds in intrinsic, i.e. inevitable and irreducible, uncertainties concerning the process of modelling as well. Therefore, it is indispensable to quantify uncertainty in order to determine, which are robust results under this inherent uncertainty. The central goal of this thesis is to explore how uncertainties map on the properties of interest such as phase space topology and qualitative dynamics of the system. We will address several types of uncertainty and apply methods of dynamical systems theory on a trendsetting field of climate research, i.e. the Indian monsoon. For the systematic analysis concerning the different facets of uncertainty, a box model of the Indian monsoon is investigated, which shows a saddle node bifurcation against those parameters that influence the heat budget of the system and that goes along with a regime shift from a wet to a dry summer monsoon. As some of these parameters are crucially influenced by anthropogenic perturbations, the question is whether the occurrence of this bifurcation is robust against uncertainties in parameters and in the number of considered processes and secondly, whether the bifurcation can be reached under climate change. Results indicate, for example, the robustness of the bifurcation point against all considered parameter uncertainties. The possibility of reaching the critical point under climate change seems rather improbable. A novel method is applied for the analysis of the occurrence and the position of the bifurcation point in the monsoon model against parameter uncertainties. This method combines two standard approaches: a bifurcation analysis with multi-parameter ensemble simulations. As a model-independent and therefore universal procedure, this method allows investigating the uncertainty referring to a bifurcation in a high dimensional parameter space in many other models. With the monsoon model the uncertainty about the external influence of El Niño / Southern Oscillation (ENSO) is determined. There is evidence that ENSO influences the variability of the Indian monsoon, but the underlying physical mechanism is discussed controversially. As a contribution to the debate three different hypotheses are tested of how ENSO and the Indian summer monsoon are linked. In this thesis the coupling through the trade winds is identified as key in linking these two key climate constituents. On the basis of this physical mechanism the observed monsoon rainfall data can be reproduced to a great extent. Moreover, this mechanism can be identified in two general circulation models (GCMs) for the present day situation and for future projections under climate change. Furthermore, uncertainties in the process of coupling models are investigated, where the focus is on a comparison of forced dynamics as opposed to fully coupled dynamics. The former describes a particular type of coupling, where the dynamics from one sub-module is substituted by data. Intrinsic uncertainties and constraints are identified that prevent the consistency of a forced model with its fully coupled counterpart. Qualitative discrepancies between the two modelling approaches are highlighted, which lead to an overestimation of predictability and produce artificial predictability in the forced system. The results suggest that bistability and intermittent predictability, when found in a forced model set-up, should always be cross-validated with alternative coupling designs before being taken for granted. All in this, this thesis contributes to the fundamental issue of dealing with uncertainties the climate modelling community is confronted with. Although some uncertainties allow for including them in the interpretation of the model results, intrinsic uncertainties could be identified, which are inevitable within a certain modelling paradigm and are provoked by the specific modelling approach.
This Thesis was devoted to the study of the coupled system composed by El Niño/Southern Oscillation and the Annual Cycle. More precisely, the work was focused on two main problems: 1. How to separate both oscillations into an affordable model for understanding the behaviour of the whole system. 2. How to model the system in order to achieve a better understanding of the interaction, as well as to predict future states of the system. We focused our efforts in the Sea Surface Temperature equations, considering that atmospheric effects were secondary to the ocean dynamics. The results found may be summarised as follows: 1. Linear methods are not suitable for characterising the dimensionality of the sea surface temperature in the tropical Pacific Ocean. Therefore they do not help to separate the oscillations by themselves. Instead, nonlinear methods of dimensionality reduction are proven to be better in defining a lower limit for the dimensionality of the system as well as in explaining the statistical results in a more physical way [1]. In particular, Isomap, a nonlinear modification of Multidimensional Scaling methods, provides a physically appealing method of decomposing the data, as it substitutes the euclidean distances in the manifold by an approximation of the geodesic distances. We expect that this method could be successfully applied to other oscillatory extended systems and, in particular, to meteorological systems. 2. A three dimensional dynamical system could be modeled, using a backfitting algorithm, for describing the dynamics of the sea surface temperature in the tropical Pacific Ocean. We observed that, although there were few data points available, we could predict future behaviours of the coupled ENSO-Annual Cycle system with an accuracy of less than six months, although the constructed system presented several drawbacks: few data points to input in the backfitting algorithm, untrained model, lack of forcing with external data and simplification using a close system. Anyway, ensemble prediction techniques showed that the prediction skills of the three dimensional time series were as good as those found in much more complex models. This suggests that the climatological system in the tropics is mainly explained by ocean dynamics, while the atmosphere plays a secondary role in the physics of the process. Relevant predictions for short lead times can be made using a low dimensional system, despite its simplicity. The analysis of the SST data suggests that nonlinear interaction between the oscillations is small, and that noise plays a secondary role in the fundamental dynamics of the oscillations [2]. A global view of the work shows a general procedure to face modeling of climatological systems. First, we should find a suitable method of either linear or nonlinear dimensionality reduction. Then, low dimensional time series could be extracted out of the method applied. Finally, a low dimensional model could be found using a backfitting algorithm in order to predict future states of the system.
When Galactic microlensing events of stars are observed, one usually measures a symmetric light curve corresponding to a single lens, or an asymmetric light curve, often with caustic crossings, in the case of a binary lens system. In principle, the fraction of binary stars at a certain separation range can be estimated based on the number of measured microlensing events. However, a binary system may produce a light curve which can be fitted well as a single lens light curve, in particullary if the data sampling is poor and the errorbars are large. We investigate what fraction of microlensing events produced by binary stars for different separations may be well fitted by and hence misinterpreted as single lens events for various observational conditions. We find that this fraction strongly depends on the separation of the binary components, reaching its minimum at between 0.6 and 1.0 Einstein radius, where it is still of the order of 5% The Einstein radius is corresponding to few A.U. for typical Galactic microlensing scenarios. The rate for misinterpretation is higher for short microlensing events lasting up to few months and events with smaller maximum amplification. For fixed separation it increases for binaries with more extreme mass ratios. Problem of degeneracy in photometric light curve solution between binary lens and binary source microlensing events was studied on simulated data, and data observed by the PLANET collaboration. The fitting code BISCO using the PIKAIA genetic algorithm optimizing routine was written for optimizing binary-source microlensing light curves observed at different sites, in I, R and V photometric bands. Tests on simulated microlensing light curves show that BISCO is successful in finding the solution to a binary-source event in a very wide parameter space. Flux ratio method is suggested in this work for breaking degeneracy between binary-lens and binary-source photometric light curves. Models show that only a few additional data points in photometric V band, together with a full light curve in I band, will enable breaking the degeneracy. Very good data quality and dense data sampling, combined with accurate binary lens and binary source modeling, yielded the discovery of the lowest-mass planet discovered outside of the Solar System so far, OGLE-2005-BLG-390Lb, having only 5.5 Earth masses. This was the first observed microlensing event in which the degeneracy between a planetary binary-lens and an extreme flux ratio binary-source model has been successfully broken. For events OGLE-2003-BLG-222 and OGLE-2004-BLG-347, the degeneracy was encountered despite of very dense data sampling. From light curve modeling and stellar evolution theory, there was a slight preference to explain OGLE-2003-BLG-222 as a binary source event, and OGLE-2004-BLG-347 as a binary lens event. However, without spectra, this degeneracy cannot be fully broken. No planet was found so far around a white dwarf, though it is believed that Jovian planets should survive the late stages of stellar evolution, and that white dwarfs will retain planetary systems in wide orbits. We want to perform high precision astrometric observations of nearby white dwarfs in wide binary systems with red dwarfs in order to find planets around white dwarfs. We selected a sample of observing targets (WD-RD binary systems, not published yet), which can possibly have planets around the WD component, and modeled synthetic astrometric orbits which can be observed for these targets using existing and future astrometric facilities. Modeling was performed for the astrometric accuracy of 0.01, 0.1, and 1.0 mas, separation between WD and planet of 3 and 5 A.U., binary system separation of 30 A.U., planet masses of 10 Earth masses, 1 and 10 Jupiter masses, WD mass of 0.5M and 1.0 Solar masses, and distances to the system of 10, 20 and 30 pc. It was found that the PRIMA facility at the VLTI will be able to detect planets around white dwarfs once it is operating, by measuring the astrometric wobble of the WD due to a planet companion, down to 1 Jupiter mass. We show for the simulated observations that it is possible to model the orbits and find the parameters describing the potential planetary systems.
Since their discovery in 1610 by Galileo Galilei, Saturn's rings continue to fascinate both experts and amateurs. Countless numbers of icy grains in almost Keplerian orbits reveal a wealth of structures such as ringlets, voids and gaps, wakes and waves, and many more. Grains are found to increase in size with increasing radial distance to Saturn. Recently discovered "propeller" structures in the Cassini spacecraft data, provide evidence for the existence of embedded moonlets. In the wake of these findings, the discussion resumes about origin and evolution of planetary rings, and growth processes in tidal environments. In this thesis, a contact model for binary adhesive, viscoelastic collisions is developed that accounts for agglomeration as well as restitution. Collisional outcomes are crucially determined by the impact speed and masses of the collision partners and yield a maximal impact velocity at which agglomeration still occurs. Based on the latter, a self-consistent kinetic concept is proposed. The model considers all possible collisional outcomes as there are coagulation, restitution, and fragmentation. Emphasizing the evolution of the mass spectrum and furthermore concentrating on coagulation alone, a coagulation equation, including a restricted sticking probability is derived. The otherwise phenomenological Smoluchowski equation is reproduced from basic principles and denotes a limit case to the derived coagulation equation. Qualitative and quantitative analysis of the relevance of adhesion to force-free granular gases and to those under the influence of Keplerian shear is investigated. Capture probability, agglomerate stability, and the mass spectrum evolution are investigated in the context of adhesive interactions. A size dependent radial limit distance from the central planet is obtained refining the Roche criterion. Furthermore, capture probability in the presence of adhesion is generally different compared to the case of pure gravitational capture. In contrast to a Smoluchowski-type evolution of the mass spectrum, numerical simulations of the obtained coagulation equation revealed, that a transition from smaller grains to larger bodies cannot occur via a collisional cascade alone. For parameters used in this study, effective growth ceases at an average size of centimeters.
<img src="http://vg00.met.vgwort.de/na/806c85cec18906a64e06" width="1" height="1" alt=""> Subject of this work is the possibility to synchronize nonlinear systems via correlated noise and automatic control. The thesis is divided into two parts. The first part is motivated by field studies on feral sheep populations on two islands of the St. Kilda archipelago, which revealed strong correlations due to environmental noise. For a linear system the population correlation equals the noise correlation (Moran effect). But there exists no systematic examination of the properties of nonlinear maps under the influence of correlated noise. Therefore, in the first part of this thesis the noise-induced correlation of logistic maps is systematically examined. For small noise intensities it can be shown analytically that the correlation of quadratic maps in the fixed-point regime is always smaller than or equal to the noise correlation. In the period-2 regime a Markov model explains qualitatively the main dynamical characteristics. Furthermore, two different mechanisms are introduced which lead to a higher correlation of the systems than the environmental correlation. The new effect of "correlation resonance" is described, i. e. the correlation yields a maximum depending on the noise intensity. In the second part of the thesis an automatic control method is presented which synchronizes different systems in a robust way. This method is inspired by phase-locked loops and is based on a feedback loop with a differential control scheme, which allows to change the phases of the controlled systems. The effectiveness of the approach is demonstrated for controlled phase synchronization of regular oscillators and foodweb models.
In this work the first observation of new type of liquid crystals is presented. This is ionic self-assembly (ISA) liquid crystals formed by introduction of oppositely charged ions between different low molecular tectonic units. As practically all conventional liquid crystals consist of rigid core and alkyl chains the attention is focused to the simplest case where oppositely charged ions are placed between a rigid core and alkyl tails. The aim of this work is to investigate and understand liquid crystalline and alignment properties of these materials. It was found that ionic interactions within complexes play the main role. Presence of these interactions restricts transition to isotropic phase. In addition, these interactions hold the system (like network) allowing crystallization into a single domain from aligned LC state. Alignment of these simple ISA complexes was spontaneous on a glass substrate. In order to show potentials for application perylenediimide and azobenzene containing ISA complexes have been investigated for correlations between phase behavior and their alignment properties. The best results of macroscopic alignment of perylenediimide-based ISA complexes have been obtained by zone-casting method. In the aligned films the columns of the complex align perpendicular to the phase-transition front. The obtained anisotropy (DR = 18) is thermally stable. The investigated photosensitive (azobenzene-based) ISA complexes show formation of columnar LC phases. It was demonstrated that photo alignment of such complexes was very effective (DR = 50 has been obtained). It was shown that photo-reorientation in the photosensitive ISA complexes is cooperative process. The size of domains has direct influence on efficiency of the photo-reorientation process. In the case of small domains the photo-alignment is the most effective. Under irradiation with linearly polarized light domains reorient in the plane of the film leading to macroscopic alignment of columns parallel to the light polarization and joining of small domains into big ones. Finally, the additional distinguishable properties of the ISA liquid crystalline complexes should be noted: (I) the complexes do not solve in water but readily solve in organic solvents; (II) the complexes have good film-forming properties when cast or spin-coated from organic solvent; (III) alignment of the complexes depends on their structure and secondary interactions between tectonic units.
In Leuchtdioden wird Licht durch die Rekombination von injizierten Ladungsträgern erzeugt. Das kann einerseits in anorganischen Materialien geschehen. In diesem Fall ist es notwendig, hochgeordnete Kristallstrukturen herzustellen, die die Eigenschaften der Leuchtdioden bestimmen. Ein anderer Ansatz ist die Verwendung von organischen Molekülen und Polymeren. Auf Grund der Vielseitigkeit der organischen Chemie können die Eigenschaften der verwendeten halbleitenden Polymere schon während der Synthese beeinflusst werden. Außerdem weisen auch diese Polymere die bekannte mechanische Flexibilität auf. Die Herstellung von flexiblen, großflächigen Beleuchtungsquellen und Anzeigelementen ist so möglich. Die erste Leuchtdiode mit einem halbleitenden Polymer als Emitter wurde 1990 hergestellt. Seither hat das Forschungsgebiet eine rasante Entwicklung genommen. Auch erste kommerzielle Produkte sind erhältlich. Im Zuge dieser Entwicklung wurde deutlich, dass die Eigenschaften von polymeren Leuchtdioden – beispielsweise Farbe und Effizienz – durch die Verwendung mehrerer Komponenten in der aktiven Schicht deutlich verbessert werden können. Gleichzeitig ergeben sich neue Herausforderungen durch die Wechselwirkungen der verschiedenen Filmbestandteile. Während die Komponenten oft entweder zur Verbesserung des Ladungstransportes oder zur Beeinflussung der Emission zugegeben werden, muss darauf geachtet werden, dass die anderen Prozesse nicht negativ beeinflusst werden. In dieser Arbeit werden einige dieser Wechselwirkungen untersucht und mit einfachen physikalischen Modellen erklärt. So werden zunächst blau emittierende Leuchtdioden auf der Basis von Polyfluoren untersucht. Dieses Material ist zwar ein sehr effizienter blauer Emitter, jedoch ist es anfällig für chemische Defekte, diese sich nicht vollständig verhindern lassen. Die Defekte bilden Fallenzustände für Elektronen, ihr Einfluss lässt sich durch die Zugabe von Lochfallen unterdrücken. Der zugrunde liegende Prozess, die Beeinflussung der Ladungsträgerbalance, wird erklärt. Im Folgenden werden Mischsystemen mit dendronisierten Emittern, die gleichzeitig eine Falle für Elektronen bilden, untersucht. Hier wird die unterschiedliche Wirkung der isolierenden Hülle auf die Ladungs- und Energieübertragung zwischen Matrix und Farbstoffkern der Dendrimere untersucht. In Mischsystemen haben die Natur der angeregten Zustände sowie die Art und Weise des Ladungsträgertransportes einen großen Einfluss auf diese Transferprozesse. Außerden hat auch hier die Ladungsträgerbalance Auswirkungen auf die Emission. Um den Ladungsträgereinfang in Fallenzuständen zu charakterisieren, wird eine Methode auf Grundlage der Messung des zeitaufgelösten Photostroms in organischen Mischfilmen weiterentwickelt. Die erzielten Ergebnisse zeigen, dass die Übertragung der für geordnete Systeme entwickelten Modelle des Ladungsträgertransportes nicht ohne weiteres auf Polymersysteme mit hoher Unordnung übertragen werden können. Abschließend werden zeitaufgelöste Messungen der Phosphoreszenz in entsprechenden Mischungen aus Polymeren und organometallischen Verbindungen vorgestellt. Auch diese Systeme enthalten üblicherweise weitere Komponenten, die den Ladungstransport verbessern. In diesen Filmen kann es zu einer Übertragung der Tripletts vom Emitter auf die weiteren Filmbestandteile kommen. Bei Kenntnis der in Frage kommenden Wechselwirkungen können die unerwünschten Prozesse vermieden werden.
In view of the importance of charge storage in polymer electrets for electromechanical transducer applications, the aim of this work is to contribute to the understanding of the charge-retention mechanisms. Furthermore, we will try to explain how the long-term storage of charge carriers in polymeric electrets works and to identify the probable trap sites. Charge trapping and de-trapping processes were investigated in order to obtain evidence of the trap sites in polymeric electrets. The charge de-trapping behavior of two particular polymer electrets was studied by means of thermal and optical techniques. In order to obtain evidence of trapping or de-trapping, charge and dipole profiles in the thickness direction were also monitored. In this work, the study was performed on polyethylene terephthalate (PETP) and on cyclic-olefin copolymers (COCs). PETP is a photo-electret and contains a net dipole moment that is located in the carbonyl group (C = O). The electret behavior of PETP arises from both the dipole orientation and the charge storage. In contrast to PETP, COCs are not photo-electrets and do not exhibit a net dipole moment. The electret behavior of COCs arises from the storage of charges only. COC samples were doped with dyes in order to probe their internal electric field. COCs show shallow charge traps at 0.6 and 0.11 eV, characteristic for thermally activated processes. In addition, deep charge traps are present at 4 eV, characteristic for optically stimulated processes. PETP films exhibit a photo-current transient with a maximum that depends on the temperature with an activation energy of 0.106 eV. The pair thermalization length (rc) calculated from this activation energy for the photo-carrier generation in PETP was estimated to be approx. 4.5 nm. The generated photo-charge carriers can recombine, interact with the trapped charge, escape through the electrodes or occupy an empty trap. PETP possesses a small quasi-static pyroelectric coefficient (QPC): ~0.6 nC/(m²K) for unpoled samples, ~60 nC/(m²K) for poled samples and ~60 nC/(m²K) for unpoled samples under an electric bias (E ~10 V/µm). When stored charges generate an internal electric field of approx. 10 V/µm, they are able to induce a QPC comparable to that of the oriented dipoles. Moreover, we observe charge-dipole interaction. Since the raw data of the QPC-experiments on PETP samples is noisy, a numerical Fourier-filtering procedure was applied. Simulations show that the data analysis is reliable when the noise level is up to 3 times larger than the calculated pyroelectric current for the QPC. PETP films revealed shallow traps at approx. 0.36 eV during thermally-stimulated current measurements. These energy traps are associated with molecular dipole relaxations (C = O). On the other hand, photo-activated measurements yield deep charge traps at 4.1 and 5.2 eV. The observed wavelengths belong to the transitions in PETP that are analogous to the π - π* benzene transitions. The observed charge de-trapping selectivity in the photocharge decay indicates that the charge detrapping is from a direct photon-charge interaction. Additionally, the charge de-trapping can be facilitated by photo-exciton generation and the interaction of the photo-excitons with trapped charge carriers. These results indicate that the benzene rings (C6H4) and the dipolar groups (C = O) can stabilize and share an extra charge carrier in a chemical resonance. In this way, this charge could be de-trapped in connection with the photo-transitions of the benzene ring and with the dipole relaxations. The thermally-activated charge release shows a difference in the trap depth to its optical counterpart. This difference indicates that the trap levels depend on the de-trapping process and on the chemical nature of the trap site. That is, the processes of charge detrapping from shallow traps are related to secondary forces. The processes of charge de-trapping from deep traps are related to primary forces. Furthermore, the presence of deep trap levels causes the stability of the charge for long periods of time.
What can we learn from climate data? : Methods for fluctuation, time/scale and phase analysis
(2006)
Since Galileo Galilei invented the first thermometer, researchers have tried to understand the complex dynamics of ocean and atmosphere by means of scientific methods. They observe nature and formulate theories about the climate system. Since some decades powerful computers are capable to simulate the past and future evolution of climate. Time series analysis tries to link the observed data to the computer models: Using statistical methods, one estimates characteristic properties of the underlying climatological processes that in turn can enter the models. The quality of an estimation is evaluated by means of error bars and significance testing. On the one hand, such a test should be capable to detect interesting features, i.e. be sensitive. On the other hand, it should be robust and sort out false positive results, i.e. be specific. This thesis mainly aims to contribute to methodological questions of time series analysis with a focus on sensitivity and specificity and to apply the investigated methods to recent climatological problems. First, the inference of long-range correlations by means of Detrended Fluctuation Analysis (DFA) is studied. It is argued that power-law scaling of the fluctuation function and thus long-memory may not be assumed a priori but have to be established. This requires to investigate the local slopes of the fluctuation function. The variability characteristic for stochastic processes is accounted for by calculating empirical confidence regions. The comparison of a long-memory with a short-memory model shows that the inference of long-range correlations from a finite amount of data by means of DFA is not specific. When aiming to infer short memory by means of DFA, a local slope larger than $\alpha=0.5$ for large scales does not necessarily imply long-memory. Also, a finite scaling of the autocorrelation function is shifted to larger scales in the fluctuation function. It turns out that long-range correlations cannot be concluded unambiguously from the DFA results for the Prague temperature data set. In the second part of the thesis, an equivalence class of nonstationary Gaussian stochastic processes is defined in the wavelet domain. These processes are characterized by means of wavelet multipliers and exhibit well defined time dependent spectral properties; they allow one to generate realizations of any nonstationary Gaussian process. The dependency of the realizations on the wavelets used for the generation is studied, bias and variance of the wavelet sample spectrum are calculated. To overcome the difficulties of multiple testing, an areawise significance test is developed and compared to the conventional pointwise test in terms of sensitivity and specificity. Applications to Climatological and Hydrological questions are presented. The thesis at hand mainly aims to contribute to methodological questions of time series analysis and to apply the investigated methods to recent climatological problems. In the last part, the coupling between El Nino/Southern Oscillation (ENSO) and the Indian Monsoon on inter-annual time scales is studied by means of Hilbert transformation and a curvature defined phase. This method allows one to investigate the relation of two oscillating systems with respect to their phases, independently of their amplitudes. The performance of the technique is evaluated using a toy model. From the data, distinct epochs are identified, especially two intervals of phase coherence, 1886-1908 and 1964-1980, confirming earlier findings from a new point of view. A significance test of high specificity corroborates these results. Also so far unknown periods of coupling invisible to linear methods are detected. These findings suggest that the decreasing correlation during the last decades might be partly inherent to the ENSO/Monsoon system. Finally, a possible interpretation of how volcanic radiative forcing could cause the coupling is outlined.
Uncertainty about the sensitivity of the climate system to changes in the Earth’s radiative balance constitutes a primary source of uncertainty for climate projections. Given the continuous increase in atmospheric greenhouse gas concentrations, constraining the uncertainty range in such type of sensitivity is of vital importance. A common measure for expressing this key characteristic for climate models is the climate sensitivity, defined as the simulated change in global-mean equilibrium temperature resulting from a doubling of atmospheric CO2 concentration. The broad range of climate sensitivity estimates (1.5-4.5°C as given in the last Assessment Report of the Intergovernmental Panel on Climate Change, 2001), inferred from comprehensive climate models, illustrates that the strength of simulated feedback mechanisms varies strongly among different models. The central goal of this thesis is to constrain uncertainty in climate sensitivity. For this objective we first generate a large ensemble of model simulations, covering different feedback strengths, and then request their consistency with present-day observational data and proxy-data from the Last Glacial Maximum (LGM). Our analyses are based on an ensemble of fully-coupled simulations, that were realized with a climate model of intermediate complexity (CLIMBER-2). These model versions cover a broad range of different climate sensitivities, ranging from 1.3 to 5.5°C, and have been generated by simultaneously perturbing a set of 11 model parameters. The analysis of the simulated model feedbacks reveals that the spread in climate sensitivity results from different realizations of the feedback strengths in water vapour, clouds, lapse rate and albedo. The calculated spread in the sum of all feedbacks spans almost the entire plausible range inferred from a sampling of more complex models. We show that the requirement for consistency between simulated pre-industrial climate and a set of seven global-mean data constraints represents a comparatively weak test for model sensitivity (the data constrain climate sensitivity to 1.3-4.9°C). Analyses of the simulated latitudinal profile and of the seasonal cycle suggest that additional present-day data constraints, based on these characteristics, do not further constrain uncertainty in climate sensitivity. The novel approach presented in this thesis consists in systematically combining a large set of LGM simulations with data information from reconstructed regional glacial cooling. Irrespective of uncertainties in model parameters and feedback strengths, the set of our model versions reveals a close link between the simulated warming due to a doubling of CO2, and the cooling obtained for the LGM. Based on this close relationship between past and future temperature evolution, we define a method (based on linear regression) that allows us to estimate robust 5-95% quantiles for climate sensitivity. We thus constrain the range of climate sensitivity to 1.3-3.5°C using proxy-data from the LGM at low and high latitudes. Uncertainties in glacial radiative forcing enlarge this estimate to 1.2-4.3°C, whereas the assumption of large structural uncertainties may increase the upper limit by an additional degree. Using proxy-based data constraints for tropical and Antarctic cooling we show that very different absolute temperature changes in high and low latitudes all yield very similar estimates of climate sensitivity. On the whole, this thesis highlights that LGM proxy-data information can offer an effective means of constraining the uncertainty range in climate sensitivity and thus underlines the potential of paleo-climatic data to reduce uncertainty in future climate projections.
The layer-by-layer assembly (LBL) of polyelectrolytes has been extensively studied for the preparation of ultrathin films due to the versatility of the build-up process. The control of the permeability of these layers is particularly important as there are potential drug delivery applications. Multilayered polyelectrolyte microcapsules are also of great interest due to their possible use as microcontainers. This work will present two methods that can be used as employable drug delivery systems, both of which can encapsulate an active molecule and tune the release properties of the active species. Poly-(N-isopropyl acrylamide), (PNIPAM) is known to be a thermo-sensitive polymer that has a Lower Critical Solution Temperature (LCST) around 32oC; above this temperature PNIPAM is insoluble in water and collapses. It is also known that with the addition of salt, the LCST decreases. This work shows Differential Scanning Calorimetry (DSC) and Confocal Laser Scanning Microscopy (CLSM) evidence that the LCST of the PNIPAM can be tuned with salt type and concentration. Microcapsules were used to encapsulate this thermo-sensitive polymer, resulting in a reversible and tunable stimuli- responsive system. The encapsulation of the PNIPAM inside of the capsule was proven with Raman spectroscopy, DSC (bulk LCST measurements), AFM (thickness change), SEM (morphology change) and CLSM (in situ LCST measurement inside of the capsules). The exploitation of the capsules as a microcontainer is advantageous not only because of the protection the capsules give to the active molecules, but also because it facilitates easier transport. The second system investigated demonstrates the ability to reduce the permeability of polyelectrolyte multilayer films by the addition of charged wax particles. The incorporation of this hydrophobic coating leads to a reduced water sensitivity particularly after heating, which melts the wax, forming a barrier layer. This conclusion was proven with Neutron Reflectivity by showing the decreased presence of D2O in planar polyelectrolyte films after annealing creating a barrier layer. The permeability of capsules could also be decreased by the addition of a wax layer. This was proved by the increase in recovery time measured by Florescence Recovery After Photobleaching, (FRAP) measurements. In general two advanced methods, potentially suitable for drug delivery systems, have been proposed. In both cases, if biocompatible elements are used to fabricate the capsule wall, these systems provide a stable method of encapsulating active molecules. Stable encapsulation coupled with the ability to tune the wall thickness gives the ability to control the release profile of the molecule of interest.
Collisions of black holes and neutron stars, named mixed binaries in the following, are interesting because of at least two reasons. Firstly, it is expected that they emit a large amount of energy as gravitational waves, which could be measured by new detectors. The form of those waves is expected to carry information about the internal structure of such systems. Secondly, collisions of such objects are the prime suspects of short gamma ray bursts. The exact mechanism for the energy emission is unknown so far. In the past, Newtonian theory of gravitation and modifications to it were often used for numerical simulations of collisions of mixed binary systems. However, near to such objects, the gravitational forces are so strong, that the use of General Relativity is necessary for accurate predictions. There are a lot of problems in general relativistic simulations. However, systems of two neutron stars and systems of two black holes have been studies extensively in the past and a lot of those problems have been solved. One of the remaining problems so far has been the use of hydrodynamic on excision boundaries. Inside excision regions, no evolution is carried out. Such regions are often used inside black holes to circumvent instabilities of the numerical methods near the singularity. Methods to handle hydrodynamics at such boundaries have been described and tests are shown in this work. One important test and the first application of those methods has been the simulation of a collapsing neutron star to a black hole. The success of these simulations and in particular the performance of the excision methods was an important step towards simulations of mixed binaries. Initial data are necessary for every numerical simulation. However, the creation of such initial data for general relativistic situations is in general very complicated. In this work it is shown how to obtain initial data for mixed binary systems using an already existing method for initial data of two black holes. These initial data have been used for evolutions of such systems and problems encountered are discussed in this work. One of the problems are instabilities due to different methods, which could be solved by dissipation of appropriate strength. Another problem is the expected drift of the black hole towards the neutron star. It is shown, that this can be solved by using special gauge conditions, which prevent the black hole from moving on the computational grid. The methods and simulations shown in this work are only the starting step for a much more detailed study of mixed binary system. Better methods, models and simulations with higher resolution and even better gauge conditions will be focus of future work. It is expected that such detailed studies can give information about the emitted gravitational waves, which is important in view of the newly built gravitational wave detectors. In addition, these simulations could give insight into the processes responsible for short gamma ray bursts.
We investigate the rotational and thermal properties of star-forming molecular clouds using hydrodynamic simulations. Stars form from molecular cloud cores by gravoturbulent fragmentation. Understanding the angular momentum and the thermal evolution of cloud cores thus plays a fundamental role in completing the theoretical picture of star formation. This is true not only for current star formation as observed in regions like the Orion nebula or the ρ-Ophiuchi molecular cloud but also for the formation of stars of the first or second generation in the universe. In this thesis we show how the angular momentum of prestellar and protostellar cores evolves and compare our results with observed quantities. The specific angular momentum of prestellar cores in our models agree remarkably well with observations of cloud cores. Some prestellar cores go into collapse to build up stars and stellar systems. The resulting protostellar objects have specific angular momenta that fall into the range of observed binaries. We find that collapse induced by gravoturbulent fragmentation is accompanied by a substantial loss of specific angular momentum. This eases the "angular momentum problem" in star formation even in the absence of magnetic fields. The distribution of stellar masses at birth (the initial mass function, IMF) is another aspect that any theory of star formation must explain. We focus on the influence of the thermodynamic properties of star-forming gas and address this issue by studying the effects of a piecewise polytropic equation of state on the formation of stellar clusters. We increase the polytropic exponent γ from a value below unity to a value above unity at a certain critical density. The change of the thermodynamic state at the critical density selects a characteristic mass scale for fragmentation, which we relate to the peak of the IMF observed in the solar neighborhood. Our investigation generally supports the idea that the distribution of stellar masses depends mainly on the thermodynamic state of the gas. A common assumption is that the chemical evolution of the star-forming gas can be decoupled from its dynamical evolution, with the former never affecting the latter. Although justified in some circumstances, this assumption is not true in every case. In particular, in low-metallicity gas the timescales for reaching the chemical equilibrium are comparable or larger than the dynamical timescales. In this thesis we take a first approach to combine a chemical network with a hydrodynamical code in order to study the influence of low levels of metal enrichment on the cooling and collapse of ionized gas in small protogalactic halos. Our initial conditions represent protogalaxies forming within a fossil HII region -- a previously ionized HII region which has not yet had time to cool and recombine. We show that in these regions, H2 is the dominant and most effective coolant, and that it is the amount of H2 formed that controls whether or not the gas can collapse and form stars. For metallicities Z <= 10<sup>-3 Zsun, metal line cooling alters the density and temperature evolution of the gas by less than 1% compared to the metal-free case at densities below 1 cm<sup>-3 and temperatures above 2000 K. We also find that an external ultraviolet background delays or suppresses the cooling and collapse of the gas regardless of whether it is metal-enriched or not. Finally, we study the dependence of this process on redshift and mass of the dark matter halo.
The occurrence of earthquakes is characterized by a high degree of spatiotemporal complexity. Although numerous patterns, e.g. fore- and aftershock sequences, are well-known, the underlying mechanisms are not observable and thus not understood. Because the recurrence times of large earthquakes are usually decades or centuries, the number of such events in corresponding data sets is too small to draw conclusions with reasonable statistical significance. Therefore, the present study combines both, numerical modeling and analysis of real data in order to unveil the relationships between physical mechanisms and observational quantities. The key hypothesis is the validity of the so-called "critical point concept" for earthquakes, which assumes large earthquakes to occur as phase transitions in a spatially extended many-particle system, similar to percolation models. New concepts are developed to detect critical states in simulated and in natural data sets. The results indicate that important features of seismicity like the frequency-size distribution and the temporal clustering of earthquakes depend on frictional and structural fault parameters. In particular, the degree of quenched spatial disorder (the "roughness") of a fault zone determines whether large earthquakes occur quasiperiodically or more clustered. This illustrates the power of numerical models in order to identify regions in parameter space, which are relevant for natural seismicity. The critical point concept is verified for both, synthetic and natural seismicity, in terms of a critical state which precedes a large earthquake: a gradual roughening of the (unobservable) stress field leads to a scale-free (observable) frequency-size distribution. Furthermore, the growth of the spatial correlation length and the acceleration of the seismic energy release prior to large events is found. The predictive power of these precursors is, however, limited. Instead of forecasting time, location, and magnitude of individual events, a contribution to a broad multiparameter approach is encouraging.
Stars are born in turbulent molecular clouds that fragment and collapse under the influence of their own gravity, forming a cluster of hundred or more stars. The star formation process is controlled by the interplay between supersonic turbulence and gravity. In this work, the properties of stellar clusters created by numerical simulations of gravoturbulent fragmentation are compared to those from observations. This includes the analysis of properties of individual protostars as well as statistical properties of the entire cluster. It is demonstrated that protostellar mass accretion is a highly dynamical and time-variant process. The peak accretion rate is reached shortly after the formation of the protostellar core. It is about one order of magnitude higher than the constant accretion rate predicted by the collapse of a classical singular isothermal sphere, in agreement with the observations. For a more reasonable comparison, the model accretion rates are converted to the observables bolometric temperature, bolometric luminosity, and envelope mass. The accretion rates from the simulations are used as input for an evolutionary scheme. The resulting distribution in the Tbol-Lbol-Menv parameter space is then compared to observational data by means of a 3D Kolmogorov-Smirnov test. The highest probability found that the distributions of model tracks and observational data points are drawn from the same population is 70%. The ratios of objects belonging to different evolutionary classes in observed star-forming clusters are compared to the temporal evolution of the gravoturbulent models in order to estimate the evolutionary stage of a cluster. While it is difficult to estimate absolute ages, the realtive numbers of young stars reveal the evolutionary status of a cluster with respect to other clusters. The sequence shows Serpens as the youngest and IC 348 as the most evolved of the investigated clusters. Finally the structures of young star clusters are investigated by applying different statistical methods like the normalised mean correlation length and the minimum spanning tree technique and by a newly defined measure for the cluster elongation. The clustering parameters of the model clusters correspond in many cases well to those from observed ones. The temporal evolution of the clustering parameters shows that the star cluster builds up from several subclusters and evolves to a more centrally concentrated cluster, while the cluster expands slower than new stars are formed.
This thesis studies strong, completely charged polyelectrolyte brushes. Extensive molecular dynamics simulations are performed on different polyelectrolyte brush systems using local compute servers and massively parallel supercomputers. The full Coulomb interaction of charged monomers, counterions, and salt ions is treated explicitly. The polymer chains are anchored by one of their ends to a uncharged planar surface. The chains are treated under good solvent conditions. Monovalent salt ions (1:1 type) are modelled same as counterions. The studies concentrate on three different brush systems at constant temperature and moderate Coulomb interaction strength (Bjerrum length equal to bond length): The first system consists of a single polyelectrolyte brush anchored with varying grafting density to a plane. Results show that chains are extended up to about 2/3 of their contour length. The brush thickness slightly grows with increasing anchoring density. This slight dependence of the brush height on grafting density is in contrast to the well known scaling result for the osmotic brush regime. That is why the result obtained by simulations has stimulated further development of theory as well as new experimental investigations on polyelectrolyte brushes. This observation can be understood on a semi-quantitative level using a simple scaling model that incorporates excluded volume effects in a free-volume formulation where an effective cross section is assigned to the polymer chain from where couterions are excluded. The resulting regime is called nonlinear osmotic brush regime. Recently this regime was also obtained in experiments. The second system studied consists of polyelectrolyte brushes with added salt in the nonlinear osmotic regime. Varying salt is an important parameter to tune the structure and properties of polyelectrolytes. Further motivation is due to a theoretical scaling prediction by Pincus for the salt dependence of brush thickness. In the high salt limit (salt concentration much larger than counterion concentration) the brush height is predicted to decrease with increasing external salt, but with a relatively weak power law showing an exponent -1/3. There is some experimental and theoretical work that confirms this prediction, but there are other results that are in contradiction. In such a situation simulations are performed to validate the theoretical prediction. The simulation result shows that brush thickness decreases with added salt, and indeed is in quite good agreement with the scaling prediction by Pincus. The relation between buffer concentration and the effective ion strength inside the brush at varying salt concentration is of interest both from theoretical as well as experimental point of view. The simulation result shows that mobile ions (counterions as well as salt) distribute nonhomogeneously inside and outside of the brush. To explain the relation between the internal ion concentration with the buffer concentration a Donnan equilibrium approach is employed. Modifying the Donnan approach by taking into account the self-volume of polyelectrolyte chains as indicated above, the simulation result can be explained using the same effective cross section for the polymer chains. The extended Donnan equilibrium relation represents a interesting theoretical prediction that should be checked by experimental data. The third system consist of two interacting polyelectrolyte brushes that are grafted to two parallel surfaces. The interactions between brushes are important, for instance, in stabilization of dispersions against flocculation. In the simulations pressure is evaluated as a function of separation D between the two grafting planes. The pressure behavior shows different regimes for decreasing separation. This behavior is in qualitative agreement with experimental data. At relatively weak compression the pressure behavior obtained in the simulation agrees with a 1/D power law predicted by scaling theory. Beyond that the present study could supply new insight for understanding the interaction between polyelectrolyte brushes.
Investigation of tropospheric arctic aerosol and mixed-phase clouds using airborne lidar technique
(2005)
An Airborne Mobile Aerosol Lidar (AMALi) was constructed and built at Alfred-Wegener-Institute for Polar and Marine Research (AWI) in Potsdam, Germany for the lower tropospheric aerosol and cloud research under tough arctic conditions. The system was successfully used during two AWI airborne field campaigns, ASTAR 2004 and SVALEX 2005, performed in vicinity of Spitsbergen in the Arctic. The novel evaluation schemes, the Two-Stream Inversion and the Iterative Airborne Inversion, were applied to the obtained lidar data. Thereby, calculation of the particle extinction and backscatter coefficient profiles with corresponding lidar ratio profiles characteristic for the arctic air was possible. The comparison of these lidar results with the results of other in-situ and remote instrumentation (ground based Koldewey Aerosol Raman Lidar (KARL), sunphotometer, radiosounding, satellite imagery) allowed to provided clean contra polluted (Arctic Haze) characteristics of the arctic aerosols. Moreover, the data interpretation by means of the ECMWF Operational Analyses and small-scale dispersion model EULAG allowed studying the effects of the Spitsbergens orography on the aerosol load in the Planetary Boundary Layer. With respect to the cloud studies a new methodology of alternated remote AMALi measurements with the airborne in-situ cloud optical and microphysical parameters measurements was proved feasible for the low density mixed-phase cloud studies. An example of such approach during observation of the natural cloud seeding (feeder-seeder phenomenon) with ice crystals precipitating into the lower supercooled stratocumulus deck were discussed in terms of the lidar signal intensity profiles and corresponding depolarisation ratio profiles. For parts of the cloud system characterised by almost negligible multiple scattering the calculation of the particle backscatter coefficient profiles was possible using the lidar ratio information obtained from the in-situ measurements in ice-crystal cloud and water cloud.
Bestimmung von Ozonabbauraten über der Arktis und Antarktis mittels Ozonsonden- und Satellitendaten
(2005)
Diese Arbeit beschäftigt sich mit der chemischen Ozonzerstörung im arktischen und antarktischen stratosphärischen Polarwirbel. Diese wird durch Abbauprodukte von anthropogen emittierten Fluorchlorkohlenwasserstoffen und Halonen, Chlor- und Bromradikale, verursacht. Studien in denen der gemessene und modellierte Ozonabbau verglichen wird zeigen, dass die Prozeße bekannt sind, der quantitative Verlauf allerdings nicht vollständig verstanden ist. Die Prozesse, die zur Ozonzerstörung führen sind in beiden Polarwirbeln ähnlich. Allerdings fällt als Konsequenz unterschiedlicher meteorologischer Bedingungen der chemische Ozonabbau im arktischen Polarwirbel weniger drastisch aus als über der Antarktis. Der arktische Polarwirbel ist im Mittel stärker dynamisch gestört als der antarktische und weist eine stärkere Jahr-zu-Jahr Variabilität auf. Das erschwert die Messung des chemischen Ozonabbaus. Zur Trennung des chemischen Ozonabbaus von der dynamischen Umverteilung des Ozons im arktischen Polarwirbel wurde die Matchmethode entwickelt. Bei dieser Methode werden Luftpakete innerhalb des Polarwirbels mehrfach beprobt, um den chemischen Anteil der Ozonänderung zu quantifizieren. Zur Identifizierung von doppelt beprobten Luftpaketen werden Trajektorien aus Windfeldern berechnet. Können zwei Messungen im Rahmen bestimmter Qualitätskriterien durch eine Trajektorie verbunden werden, kann die Ozondifferenz zwischen beiden Sondierungen berechnet und als chemischer Ozonabbau interpretiert werden. Eine solche Koinzidenz wird Match genannt. Der Matchmethode liegt ein statistischer Ansatz zugrunde, so dass eine Vielzahl solcher doppelt beprobter Luftmassen vorliegen muss, um gesicherte Aussagen über die Ozonzerstörung gewinnen zu können. So erhält man die Ozonzerstörung in einem bestimmten Zeitintervall, also Ozonabbauraten. Um die Anzahl an doppelt beprobten Luftpackten zu erhöhen wurde eine aktive Koordinierung der Ozonsondenaufstiege entwickelt. Im Rahmen dieser Arbeit wurden Matchkampagnen während des arktischen Winters 2002/2003 und zum ersten Mal während eines antarktischen Winter (2003) durchgeführt. Aus den gewonnenen Daten wurden Ozonabbauraten in beiden Polarwirbeln bestimmt. Diese Abbauraten dienen zum einen der Evaluierung von Modellen, ermöglichen aber auch den direkten Vergleich von Ozonabbauraten in beiden Polarwirbeln. Der Winter 2002/2003 war zu Beginn durch sehr tiefe Temperaturen in der mittleren und unteren Stratosphäre charakterisiert, so dass die Matchkampagne Ende November gestartet wurde. Ab Januar war der Polarwirbel zeitweise stark dynamisch gestört. Die Kampagne ging bis Mitte März. Für den Höhenbereich von 400 bis 550 K potentieller Temperatur (15-23 km) konnten Ozonabbauraten und der Verlust in der Gesamtsäule berechnet werden. Die Ozonabbauraten wurden in verschiedenen Tests auf ihre Stabilität überprüft. Der antarktische Polarwirbel war vom Beginn des Winters bis Mitte Oktober 2003 sehr kalt und stellte Ende September kurzzeitig den Rekord für die größte bisher aufgetretene Ozonloch-Fläche ein. Es konnten für den Kampagnenzeitraum, Anfang Juni bis Anfang Oktober, Ozonabbauraten im Höhenbereich von 400 bis 550 K potentieller Temperatur ermittelt werden. Der zeitliche Verlauf des Ozonabbaus war dabei auf fast allen Höhenniveaus identisch. Die Zunahme des Sonnenlichtes im Polarwirbel mit der Zeit führt zu einem starken Anwachsen der Ozonabbauraten. Ab Mitte September gingen die Ozonabbauraten auf Null zurück, da bis zu diesem Zeitpunkt das gesamte Ozon zwischen ca. 14 und 21 km zerstört wurde. Im letzten Teil der Arbeit wird ein neuer Algorithmus auf Basis der multivariaten Regression vorgestellt, mit dem Ozonabbauraten aus Ozonprofilen verschiedener Sensoren gleichzeitig berechnet werden können. Dabei können neben der Ozonabbaurate die systematischen Fehler zwischen den einzelnen Sensoren bestimmt werden. Dies wurde exemplarisch am antarktischen Winter 2003 für das 475 K potentielle Temperatur Niveau gezeigt. Neben den Ozonprofilen der Sonden wurden Daten von zwei Satellitenexperimenten verwendet. Die mit der multivariaten Matchtechnik berechneten Ozonabbauraten stimmen gut mit den Ozonabbauraten der Einzelsensor-Matchansätze überein.
Am Beispiel der Orgelpfeife wurde der Einfluss der Wandungsgeometrie des akustischen Wellenleiters auf die Schallabstrahlung untersucht. Für verschiedene Metalllegierungen wurden unterschiedliche Profile der Orgelpfeifenwandung verglichen: ein konisches Wandungsprofil mit zur Mündung hin abnehmender Wandungsstärke und ein paralleles Wandungsprofil mit konstanter Wandungsstärke. Für eine hohe statistische Sicherheit der Ergebnisse wurden sämtliche Untersuchungen an vier mal zehn Testpfeifen durchgeführt. Mit Ausnahme der beschriebenen Unterschiede sind die Pfeifen von gleichen Abmessungen und auf gleichen Klang intoniert. Die Überprüfung der Wandungseinflüsse auf den Klang besteht aus drei verschiedenen Untersuchungen: Erstens, einer subjektiven Hinterfragung der Wahrnehmbarkeit in einem Hörtest. Zweitens wurde der abgestrahlte Luftschall objektiv gemessen und das Spektrum der Pfeifen in seinen Komponenten (Teiltöne, Grundfrequenz) verglichen. Drittens wurde mit einer neuartigen Messtechnik die Oszillation des Pfeifenkörpers (ein einem akustischen Monopol entsprechendes "Atmen" des Querschnitts) untersucht. Die Ergebnisse belegen die Wahrnehmbarkeit unterschiedlicher Wandungsprofile als auch klare objektive Differenzen zwischen den emittierten Schallspektren. Ein Atmen mit guter Korrelation zur inneren Druckanregung bestätigt den Einfluss wandungsprofilabhängiger Oszillationen auf den Klang der Orgelpfeife. Schließlich wurde die Interaktion zweier in Abstand und Grundfrequenz nah beieinander liegender Orgelpfeifen überprüft. Als Ursache des dabei wahrnehmbaren Oktavsprung des Orgeltons konnte eine gegenphasiger Oszillation des Grundtons beider Pfeifen nachgewiesen werden.
Im Rahmen dieser Arbeit wurde ein besseres Verständnis der Kopplung der Troposphäre und der Stratosphäre in den mittleren und polaren Breiten der Nordhemisphäre (NH) auf Monatszeitskalen erzielt, die auf die Ausbreitung von quasi-stationären Wellen zurückzuführen ist. Der Schwerpunkt lag dabei auf den dynamisch aktiven Wintermonaten, welche die grösste Variabilität aufweisen. Die troposphärische Variabilität wird zum Grossteil durch bevorzugte Zirkulationsstrukturen, den Telekonnexionsmustern, bestimmt. Mittels einer rotierten EOF-Analyse der geopotenziellen Höhe in 500 hPa wurden die wichtigsten regionalen troposphärischen Telekonnexionsmuster der Nordhemisphäre berechnet. Diese lassen sich drei grossen geografischen Regionen zuordnen; dem nordatlantisch-europäischen Raum, Eurasien und dem pazifisch-nordamerikanischen Raum. Da es sich um die stärksten troposphärischen Variabilitätsmuster handelt, wurden sie als grundlegende troposphärische Grössen herangezogen, um dynamische Zusammenhänge zwischen der troposphärischen und der stratosphärischen Zirkulation zu untersuchen. Dabei wurde anhand von instantanen und zeitverzögerten Korrelationsanalysen der troposphärischen Muster mit stratosphärischen Variablen erstmalig gezeigt, dass unterschiedliche regionale troposphärische Telekonnexionsmuster unterschiedliche Auswirkungen auf die stratosphärische Zirkulation haben. Es ergaben sich für die pazifisch-nordamerikanischen Muster signifikante instantane Korrelationen mit quasi-barotropen Musterstrukturen und für die nordatlantisch-europäischen Muster zonalsymmetrische Ringstrukturen ab 1978 mit signifikanten Korrelationswerten über tropischen und subtropischen Breiten und inversen Korrelationswerten über polaren Gebieten. Bei einer Untersuchung des Einflusses der stratosphärischen Variabilität wurde gezeigt, dass sich die stärkste Kopplung von nordatlantisch-europäischen Telekonnexionsmustern mit der stratosphärischen Zirkulation bei einem in Richtung Europa verschobenen Polarwirbel ergibt, wodurch die signifikanten Korrelationen ab 1978 erklärt werden können. Eine zonal gemittelte und vor allem lokale Untersuchung der Wellenausbreitungsbedingungen während dieser stratosphärischen Situation zeigt, dass es zu schwächeren Windgeschwindigkeiten in der Stratosphäre im Bereich von Nordamerika und des westlichen Nordatlantiks kommt und sich dadurch die Wellenausbreitungsbedingungen in diesem geografischen Bereich für planetare Wellen verbessern. Durch die stärkere Wellenausbreitung kommt es zu einer stärkeren Wechselwirkung mit dem Polarjet, wobei dieser abgebremst wird. Diese Abbremsung führt zu einer Verstärkung der meridionalen Residualzirkulation. D. h., wenn es zu einer verstärkten Wellenanregung im Nordatlantik und über Europa kommt, ist die Reaktion der Residualzirkulation bei einem nach Europa verschobenem Polarwirbel besonders stark. Die quasi-barotropen Korrelationsstrukturen, die sich bei den pazifisch-nordamerikanischen Mustern zeigen, weisen aufgrund von abnehmenden Störungsamplituden mit zunehmender Höhe, keiner Westwärtsneigung und einem negativen Brechungsindex im Pazifik auf verschwindende Wellen hin, die als Lösung der Wellengleichung bei negativem Brechungsindex auftreten. Dies wird durch den Polarjet, der im Bereich des Pazifiks stets sehr weit in Richtung Norden verlagert ist, verursacht. Abschliessend wurde in dieser Arbeit untersucht, ob die gefundenen Zusammenhänge von nordatlantisch-europäischen Telekonnexionsmustern mit der stratosphärischen Zirkulation auch von einem Atmosphärenmodell wiedergegeben werden können. Dazu wurde ein transienter 40-Jahre-Klimalauf des ECHAM4.L39(DLR)/CHEM Modells mit möglichst realistischen Antrieben erstmalig auf die Kopplung der Troposphäre und der Stratosphäre analysiert. Dabei konnten sowohl die troposphärischen, als auch die stratosphärischen Variabilitätsmuster vom Modell simuliert werden. Allerdings zeigen sich in den stratosphärischen Mustern Phasenverschiebungen in den Wellenzahl-1-Strukturen und ihre Zeitreihen weisen keinen signifikanten Trend ab 1978 auf. Die Kopplung der nordatlantisch-europäischen Telekonnexionsmuster mit der stratosphärischen Zirkulation zeigt eine wesentlich schwächere Reaktion der meridionalen Residualzirkulation. Somit stellte sich heraus, dass insbesondere die stratosphärische Zirkulation im Modell starke Diskrepanzen zu den Beobachtungen zeigt, die wiederum Einfluss auf die Wellenausbreitungsbedingungen haben. Es wird damit deutlich, dass für eine richtige Wiedergabe der Wellenausbreitung und somit der Kopplung der Troposphäre und Stratosphäre die stratosphärische Zirkulation eine wichtige Rolle spielt.
This work explores the equilibrium structure and thermodynamic phase behavior of complexes formed by charged polymer chains (polyelectrolytes) and oppositely charged spheres (macroions). Polyelectrolyte-macroion complexes form a common pattern in soft-matter physics, chemistry and biology, and enter in numerous technological applications as well. From a fundamental point of view, such complexes are interesting in that they combine the subtle interplay between electrostatic interactions and elastic as well as entropic effects due to conformational changes of the polymer chain, giving rise to a wide range of structural properties. This forms the central theme of theoretical studies presented in this thesis, which concentrate on a number of different problems involving strongly coupled complexes, i.e. complexes that are characterized by a large adsorption energy and small chain fluctuations. In the first part, a global analysis of the structural phase behavior of a single polyelectrolyte-macroion complex is presented based on a dimensionless representation, yielding results that cover a wide range of realistic system parameters. Emphasize is made on the interplay between the effects due to the polyelectrolytes chain length, salt concentration and the macroion charge as well as the mechanical chain persistence length. The results are summarized into generic phase diagrams characterizing the wrapping-dewrapping behavior of a polyelectrolyte chain on a macroion. A fully wrapped chain state is typically obtained at intermediate salt concentrations and chain lengths, where the amount of polyelectrolyte charge adsorbed on the macroion typically exceeds the bare macroion charge leading thus to a highly overcharged complex. Perhaps the most striking features occur when a single long polyelectrolyte chain is complexed with many oppositely charged spheres. In biology, such complexes form between DNA (which carries the cell's genetic information) and small oppositely charged histone proteins serving as an efficient mechanism for packing a huge amount of DNA into the micron-size cell nucleus in eucaryotic cells. The resultant complex fiber, known as the chromatin fiber, appears with a diameter of 30~nm under physiological conditions. Recent experiments indicate a zig-zag spatial arrangement for individual DNA-histone complexes (nucleosome core particles) along the chromatin fiber. A numerical method is introduced in this thesis based on a simple generic chain-sphere cell model that enables one to investigate the mechanism of fiber formation on a systematic level by incorporating electrostatic and elastic contributions. As will be shown, stable complex fibers exhibit an impressive variety of structures including zig-zag, solenoidal and beads-on-a-string patterns, depending on system parameters such as salt concentration, sphere charge as well as the chain contour length (per sphere). The present results predict fibers of compact zig-zag structure within the physiologically relevant regime with a diameter of about 30~nm, when DNA-histone parameters are adopted. In the next part, a numerical method is developed in order to investigate the role of thermal fluctuations on the structure and thermodynamic phase behavior of polyelectrolyte-macroion complexes. This is based on a saddle-point approximation, which allows to describe the experimentally observed reaction (or complexation) equilibrium in a dilute solution of polyelectrolytes and macroions on a systematic level. This equilibrium is determined by the entropy loss a single polyelectrolyte chain suffers as it binds to an oppositely charged macroion. This latter quantity can be calculated from the spectrum of polyelectrolyte fluctuations around a macroion, which is determined by means of a normal-mode analysis. Thereby, a stability phase diagram is obtained, which exhibits qualitative agreement with experimental findings. At elevated complex concentrations, one needs to account for the inter-complex interactions as well. It will be shown that at small separations, complexes undergo structural changes in such a way that positive patches from one complex match up with negative patches on the other. Furthermore, one of the polyelectrolyte chains may bridge between the two complexes. These mechanisms lead to a strong inter-complex attraction. As a result, the second virial coefficient associated with the inter-complex interaction becomes negative at intermediate salt concentrations in qualitative agreement with recent experiments on solutions of nucleosome core particles.
Unter atmosphärischen Zirkulationsregimen versteht man bevorzugte quasi-stationäre Zustände der atmosphärischen Zirkulation auf der planetaren Skala, die für eine bis mehrere Wochen persistieren können. Klimaänderungen, ob natürlich entstanden oder anthropogen verursacht, äußern sich in erster Linie durch Änderungen der Auftrittswahrscheinlichkeiten der natürlichen Regime. In der vorliegenden Arbeit wurden dynamische Mechanismen des Regimeverhaltens und der dekadischen Klimavariabilität der Atmosphäre bei Abwesenheit zeitlich veränderlicher externer Einflussfaktoren untersucht. Das Hauptwerkzeug dafür war ein quasi-geostrophisches Dreischichtenmodell der winterlichen atmosphärischen Zirkulation auf der Nordhemisphäre, das eine spektrale T21-Auflösung, einen orographischen und einen zeitlich konstanten thermischen Antrieb mit nicht-zonalen Anteilen besitzt. Ein solches Modell vermag großskalige atmosphärische Strömungsvorgänge außerhalb der Tropen mit einiger Genauigkeit zu simulieren. Nicht berücksichtigt werden Feuchteprozesse, die Wechselwirkung der Atmosphäre mit anderen Teilen des Klimasystems sowie anthropogene Einflüsse. Für das Dreischichtenmodell wurde ein automatisiertes, iteratives Verfahren zur Anpassung des thermischen Modellantriebs neu entwickelt. Jede Iteration des Verfahrens besteht aus einer Testintegration des Modells, ihrer Auswertung, dem Vergleich der Ergebnisse mit den NCEP-NCAR-Reanalysedaten aus den Wintermonaten Dezember, Januar und Februar sowie einer auf diesem Vergleich basierenden Antriebskorrektur. Nach Konvergenz des Verfahrens stimmt das Modell sowohl bezüglich des zonal gemittelten Klimazustandes als auch bezüglich der zeitgemittelten nicht-zonalen außertropischen diabatischen Erwärmung nahezu perfekt mit den wintergemittelten Reanalysedaten überein. In einer 1000-jährigen Simulation wurden die beobachtete mittlere Zirkulation im Winter sowie ihre Variabilität realitätsnah reproduziert, insbesondere die Arktische Oszillation (AO) und ihre vertikale Ausdehnung. Der AO-Index des Modells weist deutliche dekadische Schwankungen auf, die allein durch die interne Modelldynamik bedingt sind. Darüber hinaus zeigt das Modell ein Regimeverhalten, das gut mit den Beobachtungsdaten übereintimmt. Es besitzt ein Regime, das in etwa der negativen Phase der Nordatlantischen Oszillation (NAO) entspricht und eines, das der positiven Phase der AO ähnelt. Eine weit verbreitete Hypothese ist die näherungsweise Übereinstimmung zwischen Regimen und stationären Lösungen der Bewegungsgleichungen. In der vorliegenden Arbeit wurde diese Hypothese für das Dreischichtenmodell überprüft, mit negativem Resultat. Es wurden mittels eines Funktionalminimierungsverfahrens sechs verschiedene stationäre Zustände gefunden. Diese sind allesamt durch eine äußerst unrealistische Zirkulation gekennzeichnet und sind daher weit vom Modellattraktor entfernt. Fünf der sechs Zustände zeichnen sich durch einen extrem starken subtropischen Jet in der mittleren und obereren Modellschicht aus. Da die Ursache des Regimeverhaltens des Dreischichtenmodells nach wie vor unklar war, wurde auf ein einfacheres Modell, nämlich ein barotropes Modell mit T21-Auflösung zurückgegriffen. Für die Anpassung des Oberflächenantriebs wurde eine modifizierte Form der iterativen Prozedur verwendet. Die zeitgemittelte Zirkulation des barotropen Modells stimmt sehr gut mit der zeitlich und vertikal gemittelten Zirkulation des Dreischichtenmodells überein. Das dominierende räumliche Muster der Variabilität besitzt eine AO-ähnliche Struktur. Zudem besitzt das barotrope Modell zwei Regime, die näherungsweise der positiven und negativen Phase der AO entsprechen und somit auch den Regimen des Dreischichtenmodells ähneln. Im Verlauf der Justierung des Oberflächenantriebs konnte beobachtet werden, dass die zwei Regime des barotropen Modells durch die Vereinigung zweier koexistierender Attraktoren entstanden. Der wahrscheinliche Mechanismus der Attraktorvereinigung ist eine Randkrise eines der beiden Attraktoren, gefolgt von einer explosiven Bifurkation des anderen Attraktors. Es wird die Hypothese aufgestellt, dass der beim barotropen Modell vorgefundene Mechanismus der Regimeentstehung für atmosphärische Zirkulationsmodelle mit realitätsnahem Regimeverhalten Allgemeingültigkeit besitzt. Gestützt wird die Hypothese durch vier Experimente mit dem Dreischichtenmodell, bei denen jeweils der Parameter der Bodenreibung verringert und die Antriebsanpassung wiederholt wurde. Bei diesen Experimenten erhöhte sich die Persistenz und die Separiertheit der Regime bei abnehmender Reibung drastisch und damit auch der Anteil dekadischer Zeitskalen an der Variabilität. Die Zunahme der Persistenz der Regime ist charakteristisch für die Annäherung an eine inverse innere Krise, deren Existenz aber nicht nachgewiesen werden konnte.
In the present work, we discuss two subjects related to the nonequilibrium dynamics of polymers or biological filaments adsorbed to two-dimensional substrates. The first part is dedicated to thermally activated dynamics of polymers on structured substrates in the presence or absence of a driving force. The structured substrate is represented by double-well or periodic potentials. We consider both homogeneous and point driving forces. Point-like driving forces can be realized in single molecule manipulation by atomic force microscopy tips. Uniform driving forces can be generated by hydrodynamic flow or by electric fields for charged polymers. In the second part, we consider collective filament motion in motility assays for motor proteins, where filaments glide over a motor-coated substrate. The model for the simulation of the filament dynamics contains interactive deformable filaments that move under the influence of forces from molecular motors and thermal noise. Motor tails are attached to the substrate and modeled as flexible polymers (entropic springs), motor heads perform a directed walk with a given force-velocity relation. We study the collective filament dynamics and pattern formation as a function of the motor and filament density, the force-velocity characteristics, the detachment rate of motor proteins and the filament interaction. In particular, the formation and statistics of filament patterns such as nematic ordering due to motor activity or clusters due to blocking effects are investigated. Our results are experimentally accessible and possible experimental realizations are discussed.
In der vorliegenden Arbeit werden die Eigenschaften geschlossener fluider Membranen, sogenannter Vesikeln, bei endlichen Temperaturen untersucht. Dies beinhaltet Betrachtungen zur Form freier Vesikeln, eine Untersuchung des Adhäsionsverhaltens von Vesikeln an planaren Substraten sowie eine Untersuchung der Eigenschaften fluider Vesikeln in eingeschränkten Geometrien. Diese Untersuchungen fanden mit Hilfe von Monte-Carlo-Simulationen einer triangulierten Vesikeloberfläche statt. Die statistischen Eigenschaften der fluktuierenden fluiden Vesikeln wurden zum Teil mittels Freier-Energie-Profile analysiert. In diesem Zusammenhang wurde eine neuartige Histogrammethode entwickelt.<BR> Die Form für eine freie fluide Vesikel mit frei veränderlichem Volumen, die das Konfigurationsenergie-Funktional minimiert, ist im Falle verschwindender Temperatur eine Kugel. Mit Hilfe von Monte-Carlo-Simulationen sowie einem analytisch behandelbaren Modellsystem konnte gezeigt werden, daß sich dieses Ergebnis nicht auf endliche Temperaturen verallgemeinern lässt und statt dessen leicht prolate und oblate Vesikelformen gegenüber der Kugelgestalt überwiegen. Dabei ist die Wahrscheinlichkeit für eine prolate Form ein wenig gröoßer als für eine oblate. Diese spontane Asphärizität ist entropischen Ursprungs und tritt nicht bei zweidimensionalen Vesikeln auf. Durch osmotische Drücke in der Vesikel, die größer sind als in der umgebenden Flüssigkeit, lässt sich die Asphärizität reduzieren oder sogar kompensieren. Die Übergänge zwischen den beobachteten prolaten und oblaten Formen erfolgen im Bereich von Millisekunden in Abwesenheit osmotisch aktiver Partikel. Bei Vorhandensein derartiger Partikel ergeben sich Übergangszeiten im Bereich von Sekunden. Im Rahmen der Untersuchung des Adhäsionsverhaltens fluider Vesikeln an planaren, homogenen Substraten konnte mit Hilfe von Monte-Carlo-Simulationen festgestellt werden, dass die Eigenschaften der Kontaktfläche der Vesikeln stark davon abhängen, welche Kräfte den Kontakt bewirken. Für eine dominierende attraktive Wechselwirkung zwischen Substrat und Vesikelmembran sowie im Falle eines Massendichteunterschieds der Flüssigkeiten innerhalb und außerhalb der Vesikel, der die Vesikel auf das Substrat sinken lässt, fndet man innerhalb der Kontakt ache eine ortsunabhangige Verteilung des Abstands zwischen Vesikelmembran und Substrat. Drückt die Vesikel ohne Berücksichtigung osmotischer Effekte auf Grund einer Differenz der Massendichten der Membran und der umgebenden Flüssigkeit gegen das Substrat, so erhält man eine Abstandsverteilung zwischen Vesikelmembran und Substrat, die mit dem Abstand vom Rand der Kontaktfläche variiert. Dieser Effekt ist zudem temperaturabhängig. Ferner wurde die Adhäsion fluider Vesikeln an chemisch strukturierten planaren Substraten untersucht. Durch das Wechselspiel von entropischen Effekten und Konfigurationsenergien entsteht eine komplexe Abhängigkeit der Vesikelform von Biegesteifigkeit, osmotischen Bedingungen und der Geometrie der attraktiven Domänen. Für die Bestimmung der Biegesteifigkeit der Vesikelmembranen liefern die existierenden Verfahren stark voneinander abweichende Ergebnisse. In der vorliegenden Arbeit konnte mittels Monte-Carlo-Simulationen zur Bestimmung der Biegesteifigkeit anhand des Mikropipettenverfahrens von Evans gezeigt werden, dass dieses Verfahren die a priori für die Simulation vorgegebene Biegesteifigkeit im wesentlichen reproduzieren kann. Im Hinblick auf medizinisch-pharmazeutische Anwendungen ist der Durchgang fluider Vesikeln durch enge Poren relevant. In Monte-Carlo-Simulationen konnte gezeigt werden, dass ein spontaner Transport der Vesikel durch ein Konzentrationsgefälle osmotisch aktiver Substanzen, das den physiologischen Bedingungen entspricht, induziert werden kann. Es konnten die hierfür notwendigen osmotischen Bedingungen sowie die charakteristischen Zeitskalen abgeschätzt werden. Im realen Experiment sind Eindringzeiten in eine enge Pore im Bereich weniger Minuten zu erwarten. Ferner konnte beobachtet werden, dass bei Vesikeln mit einer homogenen, positiven spontanen Krümmung Deformationen hin zu prolaten Formen leichter erfolgen als bei Vesikeln ohne spontane Krümmung. Mit diesem Effekt ist eine Verringerung der Energiebarriere für das Eindringen in eine Pore verbunden, deren Radius nur wenig kleiner als der Vesikelradius ist.
The Thesis is focused on the properties of self-organized nanostructures. Atomic and electronic properties of different systems have been investigated using methods of electron diffraction, scanning tunneling microscopy and photoelectron spectroscopy. Implementation of the STM technique (including design, construction, and tuning of the UHV experimental set-up) has been done in the framework of present work. This time-consuming work is reported to greater detail in the experimental part of this Thesis. The scientific part starts from the study of quantum-size effects in the electronic structure of a two-dimensional Ag film on the supporting substrate Ni(111). Distinct quantum well states in the sp-band of Ag were observed in photoelectron spectra. Analysis of thickness- and angle-dependent photoemission supplies novel information on the properties of the interface. For the first time the Ni(111) relative band gap was indirectly probed in the ground-state through the electronic structure of quantum well states in the adlayer. This is particularly important for Ni where valence electrons are strongly correlated. Comparison of the experiment with calculations performed in the formalism of the extended phase accumulation model gives the substrate gap which is fully consistent with the one obtained by ab-initio LDA calculations. It is, however, in controversy to the band structure of Ni measured directly by photoemission. These results lend credit to the simplest view of photoemission from Ni, assigning early observed contradictions between theory and experiments to electron correlation effects in the final state of photoemission. Further, nanosystems of lower dimensionality have been studied. Stepped surfaces W(331) and W(551) were used as one-dimensional model systems and as templates for self-organization of Au nanoclusters. Photon energy dependent photoemission revealed a surface resonance which was never observed before on W(110) which is the base plane of the terrace microsurfaces. The dispersion E(k) of this state measured on stepped W(331) and W(551) with angle-resolved photoelectron spectroscopy is modified by a strong umklapp effect. It appears as two parabolas shifted symmetrically relative to the microsurface normal by half of the Brillouin zone of the step superlattice. The reported results are very important for understanding of the electronic properties of low-dimensional nanostructures. It was also established that W(331) and W(551) can serve as templates for self-organization of metallic nanostructures. A combined study of electronic and atomic properties of sub-monolayer amounts of gold deposited on these templates have shown that if the substrate is slightly pre-oxidized and the temperature is elevated, then Au can alloy with the first monolayer of W. As a result, a nanostructure of uniform clusters of a surface alloy is produced all over the steps. Such clusters feature a novel sp-band in the vicinity of the Fermi level, which appears split into constant energy levels due to effects of lateral quantization. The last and main part of this work is devoted to large-scale reconstructions on surfaces and nanostructures self-assembled on top. The two-dimensional surface carbide W(110)/C-R(15x3) has been extensively investigated. Photoemission studies of quantum size effects in the electronic structure of this reconstruction, combined with an investigation of its surface geometry, lead to an advanced structural model of the carbide overlayer. It was discovered that W(110)/C-R(15x3) can control self-organization of adlayers into nanostructures with extremely different electronic and structural properties. Thus, it was established that at elevated temperature the R(15x3) superstructure controls the self-assembly of sub-monolayer amounts of Au into nm-wide nanostripes. Based on the results of core level photoemission, the R(15x3)-induced surface alloying which takes place between Au and W can be claimed as driving force of self-organization. The observed stripes exhibit a characteristic one-dimensional electronic structure with laterally quantized d-bands. Obviously, these are very important for applications, since dimensions of electronic devices have already stepped into the nm-range, where quantum-size phenomena must undoubtedly be considered. Moreover, formation of perfectly uniform molecular clusters of C60 was demonstrated and described in terms of the van der Waals formalism. It is the first experimental observation of two-dimensional fullerene nanoclusters with "magic numbers". Calculations of the cluster potentials using the static approach have revealed characteristic minima in the interaction energy. They are achieved for 4 and 7 molecules per cluster. The obtained "magic numbers" and the corresponding cluster structures are fully consistent with the results of the STM measurements.
Self-assembly of polymeric building blocks is a powerful tool for the design of novel materials and structures that combine different properties and may respond to external stimuli. In the past decades, most studies were focused on the self-assembly of amphiphilic diblock copolymers in solution. The dissolution of these block copolymers in a solvent selective for one block results mostly in the formation of micelles. The micellar structure of diblock copolymers is inherently limited to a homogeneous core surrounded by a corona, which keeps the micelle in solution. Thus, for drug-delivery applications, such structures only offer a single domain (the hydrophobic inner core) for drug entrapment. Whereas multicompartment micelles composed of a water-soluble shell and a segregated hydrophobic core are novel, interesting morphologies for applications in a variety of fields including medicine, pharmacy and biotechnology. The separated incompatible compartments of the hydrophobic core could enable the selective entrapment and release of various hydrophobic drugs while the hydrophilic shell would permit the stabilization of these nanostructures in physiological media. However, so far, the preparation and control of stable multicompartment micellar systems are in the first stages and the number of morphological studies concerning such micelles is rather low. Thus considerably little is known about their exact inner structures. In the present study, we concentrate on four different approaches for the preparation of multicompartment micelles by self-assembly in aqueous media. A similarity of all approaches was that hydrocarbon and fluorocarbon blocks were selected for all employed copolymers since such segments tend to be strongly incompatible, and thus favor the segregation into distinct domains. Our studies have shown that the self-assembly of the utilized copolymers in aqueous solution leads in three cases to the formation of multicompartment micelles. As expected the shape and size of the micelles depend on the molecular architecture and to some extent also on the way of preparation. These novel structured colloids may serve as models as well as mimics for biological structures such as globular proteins, and may open interesting opportunities for nanotechnology applications.
Amphiphilic molecules contain a hydrophilic headgroup and a hydrophobic tail. The headgroup is polar or ionic and likes water, the tail is typically an aliphatic chain that cannot be accommodated in a polar environment. The prevailing molecular asymmetry leads to a spontaneous adsorption of amphiphiles at the air/water or oil/water interfaces. As a result, the surface tension and the surface rheology is changed. Amphiphiles are important tools to deliberately modify the interfacial properties of liquid interfaces and enable new phenomena such as foams which cannot be formed in a pure liquid. In this thesis we investigate the static and dynamic properties of adsorption layers of soluble amphiphiles at the air/water interface, the so called Gibbs monolayers. The classical way for an investigation of these systems is based on a thermodynamic analysis of the equilibrium surface tension as a function of the bulk composition in the framework of Gibbs theory. However, thermodynamics does not provide any structural information and several recent publications challenge even fundamental text book concepts. The experimental investigation faces difficulties imposed by the low surface coverage and the presence of dissolved amphiphiles in the adjacent bulk phase. In this thesis we used a suite of techniques with the sensitivity to detect less than a monolayer of molecules at the air-water interface. Some of these techniques are extremely complex such as infrared visible sum frequency generation (IR-VIS SFG) spectroscopy or second harmonic generation (SHG). Others are traditional techniques, such as ellipsometry employed in new ways and pushed to new limits. Each technique probes selectively different parts of the interface and the combination provides a profound picture of the interfacial architecture. The first part of the thesis is dedicated to the distribution of ions at interfaces. Adsorption layers of ionic amphiphiles serve as model systems allowing to produce a defined surface charge. The charge of the monolayer is compensated by the counterions. As a result of a complex zoo of interactions there will be a defined distribution of ions at the interface, however, its experimental determination is a big scientific challenge. We could demonstrate that a combination of linear and nonlinear techniques gives direct insights in the prevailing ion distribution. Our investigations reveal specific ion effects which cannot be described by classical Poisson-Boltzmann mean field type theories. Adsorption layer and bulk phase are in thermodynamic equilibrium, however, it is important to stress that there is a constant molecular exchange between adsorbed and dissolved species. This exchange process is a key element for the understanding of some of the thermodynamic properties. An excellent way to study Gibbs monolayers is to follow the relaxation from a non-equilibrium to an equilibrium state. Upon compression amphiphiles must leave the adsorption layer and dissolve in the adjacent bulk phase. Upon expansion amphiphiles must adsorb at the interface to restore the equilibrium coverage. Obviously the frequency of the expansion and compression cycles must match the molecular exchange processes. At too low frequencies the equilibrium is maintained at all times. If the frequency is too fast the system behaves as a monolayer of insoluble surfactants. In this thesis we describe an unique variant of an oscillating bubble technique that measures precisely the real and imaginary part of the complex dilational modulus E in a frequency range up to 500 Hz. The extension of about two decades in the time domain in comparison to the conventional method of an oscillating drop is a tremendous achievement. The imaginary part of the complex dilational modulus E is a consequence of a dissipative process which is interpreted as an intrinsic surface dilational viscosity. The IR-VIS SFG spectra of the interfacial water provide a molecular interpretation of the underlying dissipative process.
Electrets are materials capable of storing oriented dipoles or an electric surplus charge for long periods of time. The term "electret" was coined by Oliver Heaviside in analogy to the well-known word "magnet". Initially regarded as a mere scientific curiosity, electrets became increasingly imporant for applications during the second half of the 20th century. The most famous example is the electret condenser microphone, developed in 1962 by Sessler and West. Today, these devices are produced in annual quantities of more than 1 billion, and have become indispensable in modern communications technology. Even though space-charge electrets are widely used in transducer applications, relatively little was known about the microscopic mechanisms of charge storage. It was generally accepted that the surplus charges are stored in some form of physical or chemical traps. However, trap depths of less than 2 eV, obtained via thermally stimulated discharge experiments, conflicted with the observed lifetimes (extrapolations of experimental data yielded more than 100000 years). Using a combination of photostimulated discharge spectroscopy and simultaneous depth-profiling of the space-charge density, the present work shows for the first time that at least part of the space charge in, e.g., polytetrafluoroethylene, polypropylene and polyethylene terephthalate is stored in traps with depths of up to 6 eV, indicating major local structural changes. Based on this information, more efficient charge-storing materials could be developed in the future. The new experimental results could only be obtained after several techniques for characterizing the electrical, electromechanical and electrical properties of electrets had been enhanced with in situ capability. For instance, real-time information on space-charge depth-profiles were obtained by subjecting a polymer film to short laser-induced heat pulses. The high data acquisition speed of this technique also allowed the three-dimensional mapping of polarization and space-charge distributions. A highly active field of research is the development of piezoelectric sensor films from electret polymer foams. These materials store charges on the inner surfaces of the voids after having been subjected to a corona discharge, and exhibit piezoelectric properties far superior to those of traditional ferroelectric polymers. By means of dielectric resonance spectroscopy, polypropylene foams (presently the most widely used ferroelectret) were studied with respect to their thermal and UV stability. Their limited thermal stability renders them unsuitable for applications above 50 °C. Using a solvent-based foaming technique, we found an alternative material based on amorphous Teflon® AF, which exhibits a stable piezoelectric coefficient of 600 pC/N at temperatures up to 120 °C.
We present an application of imprecise probability theory to the quantification of uncertainty in the integrated assessment of climate change. Our work is motivated by the fact that uncertainty about climate change is pervasive, and therefore requires a thorough treatment in the integrated assessment process. Classical probability theory faces some severe difficulties in this respect, since it cannot capture very poor states of information in a satisfactory manner. A more general framework is provided by imprecise probability theory, which offers a similarly firm evidential and behavioural foundation, while at the same time allowing to capture more diverse states of information. An imprecise probability describes the information in terms of lower and upper bounds on probability. For the purpose of our imprecise probability analysis, we construct a diffusion ocean energy balance climate model that parameterises the global mean temperature response to secular trends in the radiative forcing in terms of climate sensitivity and effective vertical ocean heat diffusivity. We compare the model behaviour to the 20th century temperature record in order to derive a likelihood function for these two parameters and the forcing strength of anthropogenic sulphate aerosols. Results show a strong positive correlation between climate sensitivity and ocean heat diffusivity, and between climate sensitivity and absolute strength of the sulphate forcing. We identify two suitable imprecise probability classes for an efficient representation of the uncertainty about the climate model parameters and provide an algorithm to construct a belief function for the prior parameter uncertainty from a set of probability constraints that can be deduced from the literature or observational data. For the purpose of updating the prior with the likelihood function, we establish a methodological framework that allows us to perform the updating procedure efficiently for two different updating rules: Dempster's rule of conditioning and the Generalised Bayes' rule. Dempster's rule yields a posterior belief function in good qualitative agreement with previous studies that tried to constrain climate sensitivity and sulphate aerosol cooling. In contrast, we are not able to produce meaningful imprecise posterior probability bounds from the application of the Generalised Bayes' Rule. We can attribute this result mainly to our choice of representing the prior uncertainty by a belief function. We project the Dempster-updated belief function for the climate model parameters onto estimates of future global mean temperature change under several emissions scenarios for the 21st century, and several long-term stabilisation policies. Within the limitations of our analysis we find that it requires a stringent stabilisation level of around 450 ppm carbon dioxide equivalent concentration to obtain a non-negligible lower probability of limiting the warming to 2 degrees Celsius. We discuss several frameworks of decision-making under ambiguity and show that they can lead to a variety of, possibly imprecise, climate policy recommendations. We find, however, that poor states of information do not necessarily impede a useful policy advice. We conclude that imprecise probabilities constitute indeed a promising candidate for the adequate treatment of uncertainty in the integrated assessment of climate change. We have constructed prior belief functions that allow much weaker assumptions on the prior state of information than a prior probability would require and, nevertheless, can be propagated through the entire assessment process. As a caveat, the updating issue needs further investigation. Belief functions constitute only a sensible choice for the prior uncertainty representation if more restrictive updating rules than the Generalised Bayes'Rule are available.
Die vorliegende Arbeit beschäftigte sich mit zwei Themengebieten. Es ging zum einen um die mechanischen Eigenschaften von Polyelektrolythohlkapseln und zum anderen um die Adhäsion von Polyelektrolythohlkapseln. Die mechanischen Eigenschaften wurden mit der AFM „colloidal probe” Technik untersucht. Dabei zeigte sich, dass die Kraftdeformationskurven für kleine Deformationen den nach der Schalentheorie vorhergesagten linearen Verlauf haben. Ebenso wurde die quadratische Abhängigkeit der Federkonstanten von der Dicke bestätigt. Für PAH/PSS findet man einen E-Modul von 0.25 GPa. Zusammen mit der Tatsache, dass die Deformationskurven unabhängig von der Geschwindigkeit sind und praktisch keine Hysterese zeigen, sowie der Möglichkeit die Kapseln plastisch zu deformieren, kann man schließen, dass das System in einem glasartigen Zustand vorliegt. <pt>Erwartungsgemäß zeigte der pH einen starken Einfluss auf die PEM. Während in einem pH-Bereich zwischen 2 und 11.5 keine morphologischen Änderungen festgestellt werden konnten, vergrößerte sich der Radius bei pH = 12 um bis zu 50 %. Diese Radienänderung war reversibel und ging einher mit einem sichtbaren Weicherwerden der Kapseln. Eine Abnahme des E-Moduls um mindestens drei Größenordungen wurde durch Kraftdeformationsmessungen bestätigt. Die Kraftdeformationskurven zeigen eine starke Hysterese. Das System befindet sich nun nicht mehr in einem glasartigen Zustand, sondern ist viskos bis gummiartig geworden. Messungen an Kapseln, die mit Glutardialdehyd behandelt wurden, zeigten, dass die Behandlung das pH-abhängige Verhalten verändert. Dies kann darauf zurückgeführt werden, dass das PAH durch den Glutardialdehyd quervernetzt wird. Bei einem hohen Quervernetzungsgrad, zeigen die Kapseln keine Änderung des mechanischen Verhaltens bei pH = 12. Schwach quervernetzte Kapseln werden immer noch signifikant weicher bei pH = 12, jedoch ändert sich der Radius nicht. Außerdem wurden Multilagenkapseln untersucht, deren Stabilität nicht auf elektrostatischen Wechselwirkungen sondern auf Wasserstoffbrückenbindungen beruhte. Diese Kapseln zeigten eine deutlich höhere Steifigkeit mit E-Modulen bis zu 1 GPa. Es wurde gefunden, dass auch dieses System für kleine Deformationen ein lineares Kraft-Deformationsverhalten zeigt, und dass die Federkonstante quadratisch von der Dicke abhängt. Die Kapseln lösen sich praktisch sofort bei pH = 6.5 auf. In der Nähe dieses pHs konnte das Abnehmen der Federkonstanten verfolgt werden. Außerdem wurde das Adhäsionsverhalten von PAH/PSS Kapseln auf mit PEI-beschichtetem Glas untersucht. Die Adhäsionsflächen waren zu einem großen Teil rund und ließen sich quantitativ auswerten. Der Adhäsionsradius nimmt mit dem Kapselradius zu und mit der Dicke ab. Das Verhalten konnte mit zwei Modellen, einem für die große und einem für die kleine Deformation beschrieben werden. Das große Deformationsmodell liefert um eine Größenordung niedrigere Adhäsionsenergien als das kleine Deformationsmodell, welches mit Werten von ‑0.2 mJ/m<sup>2 Werte in einem plausiblen Bereich liefert. Es wurde gefunden, dass bei einem Verhältnis von Dicke zu Deformation von etwa eins "buckling" auftritt. Dieser Punkt markierte zugleich den Übergang von der großen zur kleinen Deformation.
Adhesion of biological cells to their environment is mediated by two-dimensional clusters of specific adhesion molecules which are assembled in the plasma membrane of the cells. Due to the activity of the cells or external influences, these adhesion sites are usually subject to physical forces. In recent years, the influence of such forces on the stability of cellular adhesion clusters was increasingly investigated. In particular, experimental methods that were originally designed for the investigation of single bond rupture under force have been applied to investigate the rupture of adhesion clusters. The transition from single to multiple bonds, however, is not trivial and requires theoretical modelling. Rupture of biological adhesion molecules is a thermally activated, stochastic process. In this work, a stochastic model for the rupture and rebinding dynamics of clusters of parallel adhesion molecules under force is presented. In particular, the influence of (i) a constant force as it may be assumed for cellular adhesion clusters is investigated and (ii) the influence of a linearly increasing force as commonly used in experiments is considered. Special attention is paid to the force-mediated cooperativity of parallel adhesion bonds. Finally, the influence of a finite distance between receptors and ligands on the binding dynamics is investigated. Thereby, the distance can be bridged by polymeric linker molecules which tether the ligands to a substrate.
Stochastic information, to be understood as "information gained by the application of stochastic methods", is proposed as a tool in the assessment of changes in climate. This thesis aims at demonstrating that stochastic information can improve the consideration and reduction of uncertainty in the assessment of changes in climate. The thesis consists of three parts. In part one, an indicator is developed that allows the determination of the proximity to a critical threshold. In part two, the tolerable windows approach (TWA) is extended to a probabilistic TWA. In part three, an integrated assessment of changes in flooding probability due to climate change is conducted within the TWA. The thermohaline circulation (THC) is a circulation system in the North Atlantic, where the circulation may break down in a saddle-node bifurcation under the influence of climate change. Due to uncertainty in ocean models, it is currently very difficult to determine the distance of the THC to the bifurcation point. We propose a new indicator to determine the system's proximity to the bifurcation point by considering the THC as a stochastic system and using the information contained in the fluctuations of the circulation around the mean state. As the system is moved closer to the bifurcation point, the power spectrum of the overturning becomes "redder", i.e. more energy is contained in the low frequencies. Since the spectral changes are a generic property of the saddle-node bifurcation, the method is not limited to the THC, but it could also be applicable to other systems, e.g. transitions in ecosystems. In part two, a probabilistic extension to the tolerable windows approach (TWA) is developed. In the TWA, the aim is to determine the complete set of emission strategies that are compatible with so-called guardrails. Guardrails are limits to impacts of climate change or to climate change itself. Therefore, the TWA determines the "maneuvering space" humanity has, if certain impacts of climate change are to be avoided. Due to uncertainty it is not possible to definitely exclude the impacts of climate change considered, but there will always be a certain probability of violating a guardrail. Therefore the TWA is extended to a probabilistic TWA that is able to consider "probabilistic uncertainty", i.e. uncertainty that can be expressed as a probability distribution or uncertainty that arises through natural variability. As a first application, temperature guardrails are imposed, and the dependence of emission reduction strategies on probability distributions for climate sensitivities is investigated. The analysis suggests that it will be difficult to observe a temperature guardrail of 2°C with high probabilities of actually meeting the target. In part three, an integrated assessment of changes in flooding probability due to climate change is conducted. A simple hydrological model is presented, as well as a downscaling scheme that allows the reconstruction of the spatio-temporal natural variability of temperature and precipitation. These are used to determine a probabilistic climate impact response function (CIRF), a function that allows the assessment of changes in probability of certain flood events under conditions of a changed climate. The assessment of changes in flooding probability is conducted in 83 major river basins. Not all floods can be considered: Events that either happen very fast, or affect only a very small area can not be considered, but large-scale flooding due to strong longer-lasting precipitation events can be considered. Finally, the probabilistic CIRFs obtained are used to determine emission corridors, where the guardrail is a limit to the fraction of world population that is affected by a predefined shift in probability of the 50-year flood event. This latter analysis has two main results. The uncertainty about regional changes in climate is still very high, and even small amounts of further climate change may lead to large changes in flooding probability in some river systems.
Origin and symmetry of the observed global magnetic fields in galaxies are not fully understood. We intend to clarify the question of the magnetic field origin and investigate the global action of the magneto-rotational instability (MRI) in galactic disks with the help of 3D global magneto-hydrodynamical (MHD) simulations. The calculations were done with the time-stepping ZEUS 3D code using massive parallelization. The alpha-Omega dynamo is known to be one of the most efficient mechanisms to reproduce the observed global galactic fields. The presence of strong turbulence is a pre-requisite for the alpha-Omega dynamo generation of the regular magnetic fields. The observed magnitude and spatial distribution of turbulence in galaxies present unsolved problems to theoreticians. The MRI is known to be a fast and powerful mechanism to generate MHD turbulence and to amplify magnetic fields. We find that the critical wavelength increases with the increasing of magnetic fields during the simulation, transporting the energy from critical to larger scales. The final structure, if not disrupted by supernovae explosions, is the structure of `thin layers' of thickness of about 100 pcs. An important outcome of all simulations is the magnitude of the horizontal components of the Reynolds and Maxwell stresses. The result is that the MRI-driven turbulence is magnetic-dominated: its magnetic energy exceeds the kinetic energy by a factor of 4. The Reynolds stress is small and less than 1% of the Maxwell stress. The angular momentum transport is thus completely dominated by the magnetic field fluctuations. The volume-averaged pitch angle is always negative with a magnitude of about -30. The non-saturated MRI regime is lasting sufficiently long to fill the time between the galactic encounters, independently of strength and geometry of the initial field. Therefore, we may claim the observed pitch angles can be due to MRI action in the gaseous galactic disks. The MRI is also shown to be a very fast instability with e-folding time proportional to the time of one rotation. Steep rotation curves imply a stronger growth for the magnetic energy due to MRI. The global e-folding time is from 44 Myr to 100 Myr depending on the rotation profile. Therefore, MRI can explain the existence of rather large magnetic field in very young galaxies. We also have reproduced the observed rms values of velocities in the interstellar turbulence as it was observed in NGC 1058. We have shown with the simulations that the averaged velocity dispersion of about 5 km/s is a typical number for the MRI-driven turbulence in galaxies, which agrees with observations. The dispersion increases outside of the disk plane, whereas supernovae-driven turbulence is found to be concentrated within the disk. In our simulations the velocity dispersion increases a few times with the heights. An additional support to the dynamo alpha-effect in the galaxies is the ability of the MRI to produce a mix of quadrupole and dipole symmetries from the purely vertical seed fields, so it also solves the seed-fields problem of the galactic dynamo theory. The interaction of magneto-rotational instability and random supernovae explosions remains an open question. It would be desirable to run the simulation with the supernovae explosions included. They would disrupt the calm ring structure produced by global MRI, may be even to the level when we can no longer blame MRI to be responsible for the turbulence.
Wetting and phase transitions play a very important role our daily life. Molecularly thin films of long-chain alkanes at solid/vapour interfaces (e.g. C30H62 on silicon wafers) are very good model systems for studying the relation between wetting behaviour and (bulk) phase transitions. Immediately above the bulk melting temperature the alkanes wet partially the surface (drops). In this temperature range the substrate surface is covered with a molecularly thin ordered, solid-like alkane film ("surface freezing"). Thus, the alkane melt wets its own solid only partially which is a quite rare phenomenon in nature. The thesis treats about how the alkane melt wets its own solid surface above and below the bulk melting temperature and about the corresponding melting and solidification processes. Liquid alkane drops can be undercooled to few degrees below the bulk melting temperature without immediate solidification. This undercooling behaviour is quite frequent and theoretical quite well understood. In some cases, slightly undercooled drops start to build two-dimensional solid terraces without bulk solidification. The terraces grow radially from the liquid drops on the substrate surface. They consist of few molecular layers with the thickness multiple of all-trans length of the molecule. By analyzing the terrace growth process one can find that, both below and above the melting point, the entire substrate surface is covered with a thin film of mobile alkane molecules. The presence of this film explains how the solid terrace growth is feeded: the alkane molecules flow through it from the undercooled drops to the periphery of the terrace. The study shows for the first time the coexistence of a molecularly thin film ("precursor") with partially wetting bulk phase. The formation and growth of the terraces is observed only in a small temperature interval in which the 2D nucleation of terraces is more likely than the bulk solidification. The nucleation mechanisms for 2D solidification are also analyzed in this work. More surprising is the terrace behaviour above bulk the melting temperature. The terraces can be slightly overheated before they melt. The melting does not occur all over the surface as a single event; instead small drops form at the terrace edge. Subsequently these drops move on the surface "eating" the solid terraces on their way. By this they grow in size leaving behind paths from were the material was collected. Both overheating and droplet movement can be explained by the fact that the alkane melt wets only partially its own solid. For the first time, these results explicitly confirm the supposed connection between the absence of overheating in solid and "surface melting": the solids usually start to melt without an energetic barrier from the surface at temperatures below the bulk melting point. Accordingly, the surface freezing of alkanes give rise of an energetic barrier which leads to overheating.
Subject of this work is the study of applications of the Galactic Microlensing effect, where the light of a distant star (source) is bend according to Einstein's theory of gravity by the gravitational field of intervening compact mass objects (lenses), creating multiple (however not resolvable) images of the source. Relative motion of source, observer and lens leads to a variation of deflection/magnification and thus to a time dependant observable brightness change (lightcurve), a so-called microlensing event, lasting weeks to months. The focus lies on the modeling of binary-lens events, which provide a unique tool to fully characterize the lens-source system and to detect extra-solar planets around the lens star. Making use of the ability of genetic algorithms to efficiently explore large and intricate parameter spaces in the quest for the global best solution, a modeling software (Tango) for binary lenses is developed, presented and applied to data sets from the PLANET microlensing campaign. For the event OGLE-2002-BLG-069 the 2nd ever lens mass measurement has been achieved, leading to a scenario, where a G5III Bulge giant at 9.4 kpc is lensed by an M-dwarf binary with total mass of M=0.51 solar masses at distance 2.9 kpc. Furthermore a method is presented to use the absence of planetary lightcurve signatures to constrain the abundance of extra-solar planets.
Due to its relevance for global climate, the realistic representation of the Atlantic meridional overturning circulation (AMOC) in ocean models is a key task. In recent years, two paradigms have evolved around what are its driving mechanisms: diapycnal mixing and Southern Ocean winds. This work aims at clarifying what sets the strength of the Atlantic overturning components in an ocean general circulation model and discusses the role of spatially inhomogeneous mixing, numerical diffusion and winds. Furthermore, the relation of the AMOC with a key quantity, the meridional pressure difference is analyzed. Due to the application of a very low diffusive tracer advection scheme, a realistic Atlantic overturning circulation can be obtained that is purely wind driven. On top of the winddriven circulation, changes of density gradients are caused by increasing the parameterized eddy diffusion in the North Atlantic and Southern Ocean. The linear relation between the maximum of the Atlantic overturning and the meridional pressure difference found in previous studies is confirmed and it is shown to be due to one significant pressure gradient between the average pressure over high latitude deep water formation regions and a relatively uniform pressure between 30°N and 30°S, which can directly be related to a zonal flow through geostrophy. Under constant Southern Ocean windstress forcing, a South Atlantic outflow in the range of 6-16 Sv is obtained for a large variety of experiments. Overall, the circulation is winddriven but its strength not uniquely determined by the Southern Ocean windstress. The scaling of the Atlantic overturning components is linear with the background vertical diffusivity, not confirming the 2/3 power law for one-hemisphere models without wind forcing. The pycnocline depth is constant in the coarse resolution model with large vertical grid extends. It suggests the ocean model operates like the Stommel box model with a linear relation of the pressure difference and fixed vertical scale for the volume transport. However, this seems only valid for vertical diffusivities smaller 0.4 cm²/s, when the dominant upwelling within the Atlantic occurs along the boundaries. For larger vertical diffusivities, a significant amount of interior upwelling occurs. It is further shown that any localized vertical mixing in the deep to bottom ocean cannot drive an Atlantic overturning. However, enhanced boundary mixing at thermocline depths is potentially important. The numerical diffusion is shown to have a large impact on the representation of the Atlantic overturning in the model. While the horizontal numerical diffusion tends to destabilize the Atlantic overturning the verital numerical diffusion denotes an amplifying mechanism.
Die Fusion von Membranen ist ein entscheidender Prozeß bei der Entwicklung von Zellen im Körper. Beispielsweise ist sie eine der Voraussetzungen bei der Befruchtung einer Eizelle durch ein Spermium oder für das Eindringen von Viren in eine Zelle. Membranfusion ist auch notwendig für den Stofftransport in die Zelle hinein oder aus ihr heraus. Die Membranfusion ist daher auch von praktischen Interesse auf den Gebieten der Pharmazeutik und des 'Bioengineering'. Oft muss eine Membran mit der infiziertin Zelle fusionieren, um ein Medikament an sein Zeil zu bringen. Deshalb ist ein Verständnis der Membranfusion von großem Interesse für die Entwicklung von gezielten und effizienten Methoden des 'drug delivery'. Dasselbe gilt für die gezielte Zufuhr von Genen bei der Gentherapie. Obwohl die Membranfusion schon vor nahezu 200 Jahren von dem deutschen Biologen und Mediziner Johannes Müller beobachtet wurde, liegt ein vollständiges Verständnis des Fusionsprozesses von Zellen und (Modell-) Membranen auch heute noch in weiter Ferne. Allerdings hat im letzten Jahrzehnt das Interesse für dieses Forschungsgebiet stark zugenommen. Wissenschaftler der unterschiedlichsten Disziplinen arbeiten daran, die Mechanismen der Membranfusion aufzudecken. Biologen untersuchen Proteine, die die Fusion auslösen, Chemiker entwickeln Moleküle, die die Fusion erleichtern, und Physiker versuchen die Antriebsmechanismen der Membranfusion zu verstehen. Neue Mikroskopietechniken und die hohe Rechenleistung moderner Computer helfen die molekulare und die makroskopische Welt der Membranfusion in einem Bild zusammenzufügen. Für unsere Untersuchungen haben wir Modellmembranen, die aus Lipiddoppelschichten bestehen, benutzt. Diese Membranen formen sogenannte Vesikel oder Liposomen, abgeschlossene Membrane, in denen eine bestimmte Menge an Flüssigkeit enthalten ist. Indem wir Rezeptoren in die Membran einbringen, schaffen wir funkionalisierte Vesikel, die sich differenzieren, kooperieren und selektiv reagieren können. Wir benutzen positiv geladene wasserlösliche Ionen, um Wechselwirkungen zwischen den Vesikeln zu vermitteln, und lassen die Rezeptoren und die Ionen den Fusionsprozess auslösen. Die Wechselwirkungen werden unter dem Mikroskop durch spezielle Mikromechanischn Gerätz Mikromechinerien kontrolliert. Mit Hilfe einer sehr schnellen digitale Bildaufnahmetechnik ist es uns gelungen, die Fusion unserer Modellmembranen aufzunehmen und in Echtzeit zu dokumentieren mit einer Auflösung von 50 µs. Unsere Messungen können vergleichen werden mit Computersimulationen des Fusionsprozesses. Diese Simulationen untersuchen Prozesse, die zwischen 0.1 und 1 Mikrosekunde dauern. Eine Herausforderung für die Zukunft wird es sein, die Lücke zwischen den in Experimenten (50µs) und den in Simulationen zugänglichen Zeitskalen von beiden Seiten her zu schließen.
It is known that the efficiency of organic light-emitting devices (OLEDs) is strongly influenced by the ’quality′ of the thin films [1]. On the basis of this conviction, the work presented in this thesis aimed to obtain a better understanding of the structure of organic thin films of general interest in the field of organic light emitting devices by using scanning probe microscopies (SPMs). A not yet reported crystal structure of quaterthiophene film grown on potassium hydrogen (KHP) is determined by optical measurements, a simulation program, diffraction at both normal incidence and grazing angle and AFM. The crystal cell is triclinic with parameters a = 0.721 nm, b = 0.632 nm, c = 0.956 nm and a = 91°, b = 91.4°, g = 91° [2]. The morphologies of four organic thin films deposited on gold are characterized by ultra high vacuum scanning tunneling microscopy (UHV-STM). Terraces in an hexanethiol monolayer, lamellar structures in an azobenzenethiol monolayer, rods in a a poly(paraphenylenevinylene) oligomer film and a granular morphology in an oxadiazole film are shown. The topographies of a series of poly(3,4-ethylenedioxythiophene)/poly(styrenesulfonate) (PEDOT/PSS) films deposited on indium-tin oxide (ITO) and gold obtained from dispersions with PEDOT:PSS weight ratios of 1:20, 1:6 and 1:1 are investigated by AFM. It is demonstrated that the films show the same topography on gold and on ITO. It is shown that the PEDOT films eliminate the spike features of ITO. It is reported that PEDOT 1:20 and 1:6 appear indistinguishable between each other but different from PEDOT 1:1 (the most conductive). Coupling STM and I-d measurements, a not yet reported structural model of PEDOT 1:1 on gold is obtained [3]. In this model the surface presents grains and the bulk particles/domains rich in PEDOT embedded in a PEDOT-poor matrix. The equation of conductivity is derived. A STM investigation of four PEDOT films deposited on ITO obtained from dispersions with the same PEDOT:PSS weight ratio of 1:1 is carried out [4]. The films differ either for the presence of sorbitol or for a different synthetic route (and they present different conductivities). For the first time a quantitative and qualitative correlation between the nanometer-scale morphology of PEDOT films with and without sorbitol and their conductivity is established.
It has been known for several years that under certain conditions electrons can be confined within thin layers even if these layers consist of metal and are supported by a metal substrate. In photoelectron spectra, these layers show characteristic discrete energy levels and it has turned out that these lead to large effects like the oscillatory magnetic coupling technically exploited in modern hard disk reading heads. The current work asks in how far the concepts underlying quantization in two-dimensional films can be transferred to lower dimensionality. This problem is approached by a stepwise transition from two-dimensional layers to one-dimensional nanostructures. On the one hand, these nanostructures are represented by terraces on atomically stepped surfaces, on the other hand by atom chains which are deposited onto these terraces up to complete coverage by atomically thin nanostripes. Furthermore, self organization effects are used in order to arrive at perfectly one-dimensional atomic arrangements at surfaces. Angle-resolved photoemission is particularly suited as method of investigation because is reveals the behavior of the electrons in these nanostructures in dependence of the spacial direction which distinguishes it from, e. g., scanning tunneling microscopy. With this method intense and at times surprisingly large effects of one-dimensional quantization are observed for various exemplary systems, partly for the first time. The essential role of bandgaps in the substrate known from two-dimensional systems is confirmed for nanostructures. In addition, we reveal an ambiguity without precedent in two-dimensional layers between spacial confinement of electrons on the one side and superlattice effects on the other side as well as between effects caused by the sample and by the measurement process. The latter effects are huge and can dominate the photoelectron spectra. Finally, the effects of reduced dimensionality are studied in particular for the d electrons of manganese which are additionally affected by strong correlation effects. Surprising results are also obtained here. ---------------------------- Die Links zur jeweiligen Source der im Appendix beigefügten Veröffentlichungen befinden sich auf Seite 83 des Volltextes.
The topic of synchronization forms a link between nonlinear dynamics and neuroscience. On the one hand, neurobiological research has shown that the synchronization of neuronal activity is an essential aspect of the working principle of the brain. On the other hand, recent advances in the physical theory have led to the discovery of the phenomenon of phase synchronization. A method of data analysis that is motivated by this finding - phase synchronization analysis - has already been successfully applied to empirical data. The present doctoral thesis ties up to these converging lines of research. Its subject are methodical contributions to the further development of phase synchronization analysis, as well as its application to event-related potentials, a form of EEG data that is especially important in the cognitive sciences. The methodical contributions of this work consist firstly in a number of specialized statistical tests for a difference in the synchronization strength in two different states of a system of two oscillators. Secondly, in regard of the many-channel character of EEG data an approach to multivariate phase synchronization analysis is presented. For the empirical investigation of neuronal synchronization a classic experiment on language processing was replicated, comparing the effect of a semantic violation in a sentence context with that of the manipulation of physical stimulus properties (font color). Here phase synchronization analysis detects a decrease of global synchronization for the semantic violation as well as an increase for the physical manipulation. In the latter case, by means of the multivariate analysis the global synchronization effect can be traced back to an interaction of symmetrically located brain areas.<BR> The findings presented show that the method of phase synchronization analysis motivated by physics is able to provide a relevant contribution to the investigation of event-related potentials in the cognitive sciences.
In this thesis the magnetohydrodynamic jet formation and the effects of magnetic diffusion on the formation of axisymmetric protostellar jets have been investigated in three different simulation sets. The time-dependent numerical simulations have been performed, using the magnetohydrodynamic ZEUS-3D code.
Robotic telescopes & Doppler imaging : measuring differential rotation on long-period active stars
(2004)
The sun shows a wide variety of magnetic-activity related phenomena. The magnetic field responsible for this is generated by a dynamo process which is believed to operate in the tachocline, which is located at the bottom of the convection zone. This dynamo is driven in part by differential rotation and in part by magnetic turbulences in the convection zone. The surface differential rotation, one key ingredient of dynamo theory, can be measured by tracing sunspot positions.To extend the parameter space for dynamo theories, one can extend these measurements to other stars than the sun. The primary obstacle in this endeavor is the lack of resolved surface images on other stars. This can be overcome by the Doppler imaging technique, which uses the rotation-induced Doppler-broadening of spectral lines to compute the surface distribution of a physical parameter like temperature. To obtain the surface image of a star, high-resolution spectroscopic observations, evenly distributed over one stellar rotation period are needed. This turns out to be quite complicated for long period stars. The upcoming robotic observatory STELLA addresses this problem with a dedicated scheduling routine, which is tailored for Doppler imaging targets. This will make observations for Doppler imaging not only easier, but also more efficient.As a preview of what can be done with STELLA, we present results of a Doppler imaging study of seven stars, all of which show evidence for differential rotation, but unfortunately the errors are of the same order of magnitude as the measurements due to unsatisfactory data quality, something that will not happen on STELLA. Both, cross-correlation analysis and the sheared image technique where used to double check the results if possible. For four of these stars, weak anti-solar differential rotation was found in a sense that the pole rotates faster than the equator, for the other three stars weak differential rotation in the same direction as on the sun was found.Finally, these new measurements along with other published measurements of differential rotation using Doppler imaging, were analyzed for correlations with stellar evolution, binarity, and rotation period. The total sample of stars show a significant correlation with rotation period, but if separated into antisolar and solar type behavior, only the subsample showing anti-solar differential rotation shows this correlation. Additionally, there is evidence for binary stars showing less differential rotation as single stars, as is suggested by theory. All other parameter combinations fail to deliver any results due to the still small sample of stars available.
Adherent cells constantly collect information about the mechanical properties of their extracellular environment by actively pulling on it through cell-matrix contacts, which act as mechanosensors. In recent years, the sophisticated use of elastic substrates has shown that cells respond very sensitively to changes in effective stiffness in their environment, which results in a reorganization of the cytoskeleton in response to mechanical input. We develop a theoretical model to predict cellular self-organization in soft materials on a coarse grained level. Although cell organization in principle results from complex regulatory events inside the cell, the typical response to mechanical input seems to be a simple preference for large effective stiffness, possibly because force is more efficiently generated in a stiffer environment. The term effective stiffness comprises effects of both rigidity and prestrain in the environment. This observation can be turned into an optimization principle in elasticity theory. By specifying the cellular probing force pattern and by modeling the environment as a linear elastic medium, one can predict preferred cell orientation and position. Various examples for cell organization, which are of large practical interest, are considered theoretically: cells in external strain fields and cells close to boundaries or interfaces for different sample geometries and boundary conditions. For this purpose the elastic equations are solved exactly for an infinite space, an elastic half space and the elastic sphere. The predictions of the model are in excellent agreement with experiments for fibroblast cells, both on elastic substrates and in hydrogels. Mechanically active cells like fibroblasts could also interact elastically with each other. We calculate the optimal structures on elastic substrates as a function of material properties, cell density and the geometry of cell positioning, respectively, that allows each cell to maximize the effective stiffness in its environment due to the traction of all the other cells. Finally, we apply Monte Carlo simulations to study the effect of noise on cellular structure formation. The model not only contributes to a better understanding of many physiological situations. In the future it could also be used for biomedical applications to optimize protocols for artificial tissues with respect to sample geometry, boundary condition, material properties or cell density.
This work deals with the connection between two basic phenomena in Nonlinear Dynamics: synchronization of chaotic systems and recurrences in phase space. Synchronization takes place when two or more systems adapt (synchronize) some characteristic of their respective motions, due to an interaction between the systems or to a common external forcing. The appearence of synchronized dynamics in chaotic systems is rather universal but not trivial. In some sense, the possibility that two chaotic systems synchronize is counterintuitive: chaotic systems are characterized by the sensitivity ti different initial conditions. Hence, two identical chaotic systems starting at two slightly different initial conditions evolve in a different manner, and after a certain time, they become uncorrelated. Therefore, at a first glance, it does not seem to be plausible that two chaotic systems are able to synchronize. But as we will see later, synchronization of chaotic systems has been demonstrated. On one hand it is important to investigate the conditions under which synchronization of chaotic systems occurs, and on the other hand, to develop tests for the detection of synchronization. In this work, I have concentrated on the second task for the cases of phase synchronization (PS) and generalized synchronization (GS). Several measures have been proposed so far for the detection of PS and GS. However, difficulties arise with the detection of synchronization in systems subjected to rather large amounts of noise and/or instationarities, which are common when analyzing experimental data. The new measures proposed in the course of this thesis are rather robust with respect to these effects. They hence allow to be applied to data, which have evaded synchronization analysis so far. The proposed tests for synchronization in this work are based on the fundamental property of recurrences in phase space.
Understanding stars, their magnetic activity phenomena and the underlying dynamo action is the foundation for understanding 'life, the universe and everything' - as stellar magnetic fields play a fundamental role for star and planet formation and for the terrestrial atmosphere and climate. Starspots are the fingerprints of magnetic field lines and thereby the most important sign of activity in a star's photosphere. However, they cannot be observed directly, as it is not (yet) possible to spacially resolve the surfaces of even the nearest neighbouring stars. Therefore, an indirect approach called 'Doppler imaging' is applied, which allows to reconstruct the surface spot distribution on rapidly rotating, active stars. In this work, data from 11 years of continuous spectroscopic observations of the active binary star EI Eridani are reduced and analysed. 34 Doppler maps are obtained and the problem of how to parameterise the information content of Doppler maps is discussed. Three approaches for parameter extraction are introduced and applied to all maps: average temperature, separated for several latitude bands; fractional spottedness; and, for the analysis of structural temperature distribution, longitudinal and latitudinal spot-occurrence functions. The resulting values do not show a distinct correlation with the proposed activity cycle as seen from photometric long-term observations, thereby suggesting that the photometric activity cycle is not accompanied by a spot cycle as seen on the Sun. The general morphology of the spot pattern on EI Eri remains persistent for the whole period of 11 years. In addition, a detailed parameter study is performed. Improved orbital parameters suggest that EI Eri might be complemented by a third star in a wide orbit of about 19 years. Preliminary differential rotation measurements are carried out, indicating an anti-solar orientation.
Die Frage nach der Herkunft und der dynamischen Entwicklung langlebiger kosmischer Magnetfelder ist in vielen Details noch unbeantwortet. Es besteht zwar kein Zweifel daran, dass das Magnetfeld der Erde und anderer kosmischer Objekte durch den sogenannten Dynamoeffekt verursacht werden, der genaue Mechanismus als auch die notwendigen Voraussetzungen und Randbedingungen der zugrundeliegenden Strömungen sind aber weitgehend unbekannt. Die für einen Dynamo interessanten Strömungsmuster, die im Inneren von Himmelskörpern durch Konvektion und differentielle Rotation entstehen, sind Konvektionsrollen parallel zur Rotationsachse. Auf einer Strömung mit eben solcher Geometrie, der sogenannten Roberts-Strömung, basieren die in der vorliegenden Arbeit untersuchten Dynamomodelle. Mit Methoden der nichtlinearen Dynamik wird versucht, das Systemverhalten bei Änderung der Systemparamter genauer zu charakterisieren. Die numerischen Untersuchungen beginnen mit einer Analyse der Dynamoaktivität der Roberts-Strömung in Abhängigkeit von den zwei freien Parametern in den Modellgleichungen, der magnetischen Prandtl-Zahl und der Stärke des Energieinputs. Gefunden werden verschiedene Lösungstypen die von einem stationären Magnetfeld über periodische bis zu chaotischen Zuständen reichen. Die yugrundeliegenden Symmetrien werden beschrieben und die Bifurkationen, die zum Wechsel der Lösungstypen führen, charakterisiert. Zusätzlich gibt es Bereiche bei sehr kleinen Prandtl-Zahlen, in denen überhaupt kein Dynamo existiert. Dieses Verhalten wird in der Literatur auch für viele andere numerisch ausgewertete Modelle beschrieben. Im Übergangsbereich zwischen dynamoaktivem und dynamoinaktivem Bereich wird das Auftreten einer sogenannten Blowout-Bifurkation gefunden. Desweiteren beschäftigt sich die Arbeit mit der Frage, inwiefern Helizität, also eine schraubenförmige Bewegung, der Strömung den Dynamoeffekt beeinflusst. Dazu werden ähnliche Strömungstypen verglichen, die sich hauptsächlich in ihrem Helizitätswert unterscheiden. Es wird gefunden, dass ein bestimmter Wert der Helizität nicht unterschritten werden darf, um einen stabilen Roberts-Dynamo zu erhalten.
In this thesis, dynamical structures and manifolds in closed chaotic flows will be investigated. The knowledge about the dynamical structures (and manifolds) of a system is of importance, since they provide us first information about the dynamics of the system - means, with their help we are able to characterize the flow and maybe even to forecast it`s dynamics. The visualization of such structures in closed chaotic flows is a difficult and often long-lasting process. Here, the so-called 'Leaking-method' will be introduced, in examples of simple mathematical maps as the baker- or sine-map, with which we are able to visualize subsets of the manifolds of the system`s chaotic saddle. Comparisons between the visualized manifolds and structures traced out by chemical or biological reactions superimposed on the same flow will be done in the example of a kinematic model of the Gulf Stream. It will be shown that with the help of the leaking method dynamical structures can be also visualized in environmental systems. In the example of a realistic model of the Mediterranean Sea, the leaking method will be extended to the 'exchange-method'. The exchange method allows us to characterize transport between two regions, to visualize transport routes and their exchange sets and to calculate the exchange times. Exchange times and sets will be shown and calculated for a northern and southern region in the western basin of the Mediterranean Sea. Furthermore, mixing properties in the Earth mantle will be characterized and geometrical properties of manifolds in a 3dimensional mathematical model (ABC map) will be investigated.
Recurrence plots, a rather promising tool of data analysis, have been introduced by Eckman et al. in 1987. They visualise recurrences in phase space and give an overview about the system's dynamics. Two features have made the method rather popular. Firstly they are rather simple to compute and secondly they are putatively easy to interpret. However, the straightforward interpretation of recurrence plots for some systems yields rather surprising results. For example indications of low dimensional chaos have been reported for stock marked data, based on recurrence plots. In this work we exploit recurrences or ``naturally occurring analogues'' as they were termed by E. Lorenz, to obtain three key results. One of which is that the most striking structures which are found in recurrence plots are hinged to the correlation entropy and the correlation dimension of the underlying system. Even though an eventual embedding changes the structures in recurrence plots considerably these dynamical invariants can be estimated independently of the special parameters used for the computation. The second key result is that the attractor can be reconstructed from the recurrence plot. This means that it contains all topological information of the system under question in the limit of long time series. The graphical representation of the recurrences can also help to develop new algorithms and exploit specific structures. This feature has helped to obtain the third key result of this study. Based on recurrences to points which have the same ``recurrence structure'', it is possible to generate surrogates of the system which capture all relevant dynamical characteristics, such as entropies, dimensions and characteristic frequencies of the system. These so generated surrogates are shadowed by a trajectory of the system which starts at different initial conditions than the time series in question. They can be used then to test for complex synchronisation.
One of the most striking features of ecological systems is their ability to undergo sudden outbreaks in the population numbers of one or a small number of species. The similarity of outbreak characteristics, which is exhibited in totally different and unrelated (ecological) systems naturally leads to the question whether there are universal mechanisms underlying outbreak dynamics in Ecology. It will be shown into two case studies (dynamics of phytoplankton blooms under variable nutrients supply and spread of epidemics in networks of cities) that one explanation for the regular recurrence of outbreaks stems from the interaction of the natural systems with periodical variations of their environment. Natural aquatic systems like lakes offer very good examples for the annual recurrence of outbreaks in Ecology. The idea whether chaos is responsible for the irregular heights of outbreaks is central in the domain of ecological modeling. This question is investigated in the context of phytoplankton blooms. The dynamics of epidemics in networks of cities is a problem which offers many ecological and theoretical aspects. The coupling between the cities is introduced through their sizes and gives rise to a weighted network which topology is generated from the distribution of the city sizes. We examine the dynamics in this network and classified the different possible regimes. It could be shown that a single epidemiological model can be reduced to a one-dimensional map. We analyze in this context the dynamics in networks of weighted maps. The coupling is a saturation function which possess a parameter which can be interpreted as an effective temperature for the network. This parameter allows to vary continously the network topology from global coupling to hierarchical network. We perform bifurcation analysis of the global dynamics and succeed to construct an effective theory explaining very well the behavior of the system.
This thesis analyses synchronization phenomena occurring in large ensembles of interacting oscillatory units. In particular, the effects of nonisochronicity (frequency dependence on the oscillator's amplitude) on the macroscopic transition to synchronization are studied in detail. The new phenomena found (Anomalous Synchronization) are investigated in populations of oscillators as well as between oscillator's ensembles.
Eine Nutzung der optischen Anisotropie dünner Schichten ist vor allem für die Displaytechnologie, die optische Datenspeicherung und für optische Sicherheitselemente von hoher Bedeutung. Diese Doktorarbeit befasst sich mit theoretischen und experimentellen Untersuchung von dreidimensionaler Anisotropie und dabei insbesondere mit der Untersuchung von lichtinduzierter dreidimensionaler Anisotropie in organischen dünnen Polymer-Schichten. Die gewonnenen Erkentnisse und entwickelten Methoden können wertvolle Beiträge für Optimierungsprozesse, wie bei der Kompensation der Blickwinkelabhängigkeit von Flüssigkristall-Displays, liefern. Die neue Methode der Immersions-Transmissions-Ellipsometrie (ITE) zur Untersuchung von dünneren Schichten wurde im Rahmen dieser Dissertation entwickelt. Diese Methode gestattet es, in Kombination mit konventioneller Reflexions- und Transmissionsellipsometrie, die absoluten dreidimensionalen Brechungsindices einer biaxialen Schicht zu bestimmen. Erstmals gelang es damit, das dreidimensionale Brechungsindexellipsoid von transparenten, dünneren (150 nm) Filmen hochgenau (drei Stellen hinter dem Komma) zu bestimmen. Die ITE-Methode hat demzufolge das Potential, auch bei noch dünneren Schichten mit Gewinn eingesetzt werden zu können. Die lichtinduzierte Generierung von dreidimensionaler Anisotropie wurde in dünnen Schichten von azobenzenhaltigen und zimtsäurehaltigen, amorphen und flüssig-kristallinen Homo- und Copolymeren untersucht. Erstmals wurden quantitative Untersuchungen zur Änderung von lichtinduzierten, dreidimensionalen Anisotropien in dünnen Schichten von azobenzenhaltigen und zimtsäurehaltigen Polymeren bei Tempern oberhalb der Glastemperatur durchgeführt. Bei vielen der untersuchten Polymere war die dreidimensionale Ordnung nach dem Bestrahlen mit polarisiertem Licht und anschließendem Tempern oberhalb der Glastemperatur scheinbar von der Schichtdicke abhängig. Die Ursache liegt wohl in der, mit der neuentwickelten ITE-Methode detektierten, planaren Ausgangsorientierung der aufgeschleuderten dünneren Schichten. Um Verkippungs-Gradienten in dickeren Polymerschichten in ihrem Verlauf zu bestimmen, wurde eine spezielle Methode unter Benutzung der Wellenleitermoden-Spektroskopie entwickelt. Quantenchemisch bestimmte, maximal induzierbare Doppelbrechungen in flüssig-kristallinen Polymeren wurden mit den experimentell gefundenen Ordnungen verglichen.
We calculate the additional carbon emissions as a result of the conversion of natural land in a process of urbanisation; and the change of carbon flows by “urbanised” ecosystems, when the atmospheric carbon is exported to the neighboring territories, from 1980 till 2050 for the eight regions of the world. As a scenario we use combined UN and demographic model′s prognoses for regional total and urban population growth. The calculations of urban areas dynamics are based on two models: the regression model and the Gamma-model. The urbanised area is sub-divided on built-up, „green“ (parks, etc.) and informal settlements (favelas) areas. The next step is to calculate the regional and world dynamics of carbon emission and export, and the annual total carbon balance. Both models give similar results with some quantitative differences. In the first model, the world annual emissions attain a maximum of 205 MtC/year between 2020-2030. Emissions will then slowly decrease. The maximum contributions are given by China and the Asia and Pacific regions. In the second model, world annual emissions increase to 1.25 GtC in 2005, beginning to decrease afterwards. If we compare the emission maximum with the annual emission caused by deforestation, 1.36GtC per year, then we can say that the role of urbanised territories (UT) is of a comparable magnitude. Regarding the world annual export of carbon by UT, we observe its monotonous growth by three times, from 24 MtC to 66 MtC in the first model, and from 249 MtC to 505 MtC in the second one. The latter, is therefore comparable to the amount of carbon transported by rivers into the ocean (196-537 MtC). By estimating the total balance we find that urbanisation shifts the total balance towards a “sink” state. The urbanisation is inhibited in the interval 2020-2030, and by 2050 the growth of urbanised areas would almost stop. Hence, the total emission of natural carbon at that stage will stabilise at the level of the 1980s (80 MtC per year). As estimated by the second model, the total balance, being almost constant until 2000, then starts to decrease at an almost constant rate. We can say that by the end of the XXI century, the total carbon balance will be equal to zero, when the exchange flows are fully balanced, and may even be negative, when the system begins to take up carbon from the atmosphere, i.e., becomes a “sink”.
My thesis is concerned with several new noise-induced phenomena in excitable neural models, especially those with FitzHugh-Nagumo dynamics. In these effects the fluctuations intrinsically present in any complex neural network play a constructive role and improve functionality. I report the occurrence of Vibrational Resonance in excitable systems. Both in an excitable electronic circuit and in the FitzHugh-Nagumo model, I show that an optimal amplitude of high-frequency driving enhances the response of an excitable system to a low-frequency signal. Additionally, the influence of additive noise and the interplay between Stochastic and Vibrational Resonance is analyzed. Further, I study systems which combine both oscillatory and excitable properties, and hence intrinsically possess two internal frequencies. I show that in such a system the effect of Stochastic Resonance can be amplified by an additional high-frequency signal which is in resonance with the oscillatory frequency. This amplification needs much lower noise intensities than for conventional Stochastic Resonance in excitable systems. I study frequency selectivity in noise-induced subthreshold signal processing in a system with many noise-supported stochastic attractors. I show that the response of the coupled elements at different noise levels can be significantly enhanced or reduced by forcing some elements into resonance with these new frequencies which correspond to appropriate phase-relations. A noise-induced phase transition to excitability is reported in oscillatory media with FitzHugh-Nagumo dynamics. This transition takes place via noise-induced stabilization of a deterministically unstable fixed point of the local dynamics, while the overall phase-space structure of the system is maintained. The joint action of coupling and noise leads to a different type of phase transition and results in a stabilization of the system. The resulting noise-induced regime is shown to display properties characteristic of excitable media, such as Stochastic Resonance and wave propagation. This effect thus allows the transmission of signals through an otherwise globally oscillating medium. In particular, these theoretical findings suggest a possible mechanism for suppressing undesirable global oscillations in neural networks (which are usually characteristic of abnormal medical conditions such as Parkinson′s disease or epilepsy), using the action of noise to restore excitability, which is the normal state of neuronal ensembles.
Die Untersuchung mikrogelinster astronomischer Objekte ermöglicht es, Informationen über die Größe und Struktur dieser Objekte zu erhalten. Im ersten Teil dieser Arbeit werden die Spektren von drei gelinsten Quasare, die mit dem Potsdamer Multi Aperture Spectrophotometer (PMAS) erhalten wurden, auf Anzeichen für Mikrolensing untersucht. In den Spektren des Vierfachquasares HE 0435-1223 und des Doppelquasares HE 0047-1756 konnten Hinweise für Mikrolensing gefunden werden, während der Doppelquasar UM 673 (Q 0142--100) keine Anzeichen für Mikrolensing zeigt. Die Invertierung der Lichtkurve eines Mikrolensing-Kausik-Crossing-Ereignisses ermöglicht es, das eindimensionale Helligkeitsprofil der gelinsten Quelle zu rekonstruieren. Dies wird im zweiten Teil dieser Arbeit untersucht. Die mathematische Beschreibung dieser Aufgabe führt zu einer Volterra'schen Integralgleichung der ersten Art, deren Lösung ein schlecht gestelltes Problem ist. Zu ihrer Lösung wird in dieser Arbeit ein lokales Regularisierungsverfahren angewendet, das an die kausale Strukture der Volterra'schen Gleichung besser angepasst ist als die bisher verwendete Tikhonov-Phillips-Regularisierung. Es zeigt sich, dass mit dieser Methode eine bessere Rekonstruktion kleinerer Strukturen in der Quelle möglich ist. Weiterhin wird die Anwendbarkeit der Regularisierungsmethode auf realistische Lichtkurven mit irregulärem Sampling bzw. größeren Lücken in den Datenpunkten untersucht.
Highly collimated, high velocity streams of hot plasma – the jets – are observed as a general phenomenon being found in a variety of astrophysical objects regarding their size and energy output. Known as jet sources are protostellar objects (T Tauri stars, embedded IR sources), galactic high energy sources ("microquasars"), and active galactic nuclei (extragalactic radio sources and quasars). Within the last two decades our knowledge regarding the processes involved in astro-physical jet formation has condensed in a kind of standard model. This is the scenario of a magnetohydrodynamically accelerated and collimated jet stream launched from the innermost part of an accretion disk close to the central object. Traditionally, the problem of jet formation is divided in two categories. One is the question how to collimate and accelerate an uncollimated low velocity disk wind into a jet. The second is the question how to initiate that outflow from a disk, i.e. how to turn accretion of matter into an ejection as a disk wind. My own work is mainly related to the first question, the collimation and acceleration process. Due to the complexity of both, the physical processes believed to be responsible for the jet launching and also the spatial configuration of the physical components of the jet source, the enigma of jet formation is not yet completely understood. On the theoretical side, there has been a substantial advancement during the last decade from purely station-ary models to time-dependent simulations lead by the vast increase of computer power. Observers, on the other hand, do not yet have the instruments at hand in order to spatially resolve observe the very jet origin. It can be expected that also the next years will yield a substantial improvement on both tracks of astrophysical research. Three-dimensional magnetohydrodynamic simu-lations will improve our understanding regarding the jet-disk interrelation and the time-dependent character of jet formation, the generation of the magnetic field in the jet source, and the interaction of the jet with the ambient medium. Another step will be the combina-tion of radiation transfer computations and magnetohydrodynamic simulations providing a direct link to the observations. At the same time, a new generation of telescopes (VLT, NGST) in combination with new instrumental techniques (IR-interferometry) will lead to a "quantum leap" in jet observation, as the resolution will then be sufficient in order to zoom into the innermost region of jet formation.
In a classical context, synchronization means adjustment of rhythms of self-sustained periodic oscillators due to their weak interaction. The history of synchronization goes back to the 17th century when the famous Dutch scientist Christiaan Huygens reported on his observation of synchronization of pendulum clocks: when two such clocks were put on a common support, their pendula moved in a perfect agreement. In rigorous terms, it means that due to coupling the clocks started to oscillate with identical frequencies and tightly related phases. Being, probably, the oldest scientifically studied nonlinear effect, synchronization was understood only in 1920-ies when E. V. Appleton and B. Van der Pol systematically - theoretically and experimentally - studied synchronization of triode generators. Since that the theory was well developed and found many applications. Nowadays it is well-known that certain systems, even rather simple ones, can exhibit chaotic behaviour. It means that their rhythms are irregular, and cannot be characterized only by one frequency. However, as is shown in the Habilitation work, one can extend the notion of phase for systems of this class as well and observe their synchronization, i.e., agreement of their (still irregular!) rhythms: due to very weak interaction there appear relations between the phases and average frequencies. This effect, called phase synchronization, was later confirmed in laboratory experiments of other scientific groups. Understanding of synchronization of irregular oscillators allowed us to address important problem of data analysis: how to reveal weak interaction between the systems if we cannot influence them, but can only passively observe, measuring some signals. This situation is very often encountered in biology, where synchronization phenomena appear on every level - from cells to macroscopic physiological systems; in normal states as well as in severe pathologies. With our methods we found that cardiovascular and respiratory systems in humans can adjust their rhythms; the strength of their interaction increases with maturation. Next, we used our algorithms to analyse brain activity of Parkinsonian patients. The results of this collaborative work with neuroscientists show that different brain areas synchronize just before the onset of pathological tremor. Morevoever, we succeeded in localization of brain areas responsible for tremor generation.
One of the classical ways to describe the dynamics of nonlinear systems is to analyze theur Fourier spectra. For periodic and quasiperiodic processes the Fourier spectrum consists purely of discrete delta-functions. On the contrary, the spectrum of a chaotic motion is marked by the presence of the continuous component. In this work, we describe the peculiar, neither regular nor completely chaotic state with so called singular-continuous power spectrum. Our investigations concern various cases from most different fields, where one meets the singular continuous (fractal) spectra. The examples include both the physical processes which can be reduced to iterated discrete mappings or even symbolic sequences, and the processes whose description is based on the ordinary or partial differential equations.
Line driven winds are accelerated by the momentum transfer from photons to a plasma, by absorption and scattering in numerous spectral lines. Line driving is most efficient for ultraviolet radiation, and at plasma temperatures from 10^4 K to 10^5 K. Astronomical objects which show line driven winds include stars of spectral type O, B, and A, Wolf-Rayet stars, and accretion disks over a wide range of scales, from disks in young stellar objects and cataclysmic variables to quasar disks. It is not yet possible to solve the full wind problem numerically, and treat the combined hydrodynamics, radiative transfer, and statistical equilibrium of these flows. The emphasis in the present writing is on wind hydrodynamics, with severe simplifications in the other two areas. I consider three topics in some detail, for reasons of personal involvement. 1. Wind instability, as caused by Doppler de-shadowing of gas parcels. The instability causes the wind gas to be compressed into dense shells enclosed by strong shocks. Fast clouds occur in the space between shells, and collide with the latter. This leads to X-ray flashes which may explain the observed X-ray emission from hot stars. 2. Wind runaway, as caused by a new type of radiative waves. The runaway may explain why observed line driven winds adopt fast, critical solutions instead of shallow (or breeze) solutions. Under certain conditions the wind settles on overloaded solutions, which show a broad deceleration region and kinks in their velocity law. 3. Magnetized winds, as launched from accretion disks around stars or in active galactic nuclei. Line driving is assisted by centrifugal forces along co-rotating poloidal magnetic field lines, and by Lorentz forces due to toroidal field gradients. A vortex sheet starting at the inner disk rim can lead to highly enhanced mass loss rates.
The behaviour of an adhering cell is strongly influenced by the chemical, topographical and mechanical properties of the surface it attaches to. During recent years, it has been found experimentally that adhering cells actively sense the elastic properties of their environment by pulling on it through numerous sites of adhesion. The resulting build-up of force at sites of adhesion depends on the elastic properties of the environment and is converted into corresponding biochemical signals, which can trigger cellular programmes like growth, differentiation, apoptosis, and migration. In general, force is an important regulator of biological systems, for example in hearing and touch, in wound healing, and in rolling adhesion of leukocytes on vessel walls. In the habilitation thesis by Ulrich Schwarz, several theoretical projects are presented which address the role of forces and elasticity in cell adhesion. (1) A new method has been developed for calculating cellular forces exerted at sites of focal adhesion on micro-patterned elastic substrates. The main result is that cell-matrix contacts function as mechanosensors, converting internal force into protein aggregation. (2) A one-step master equation for the stochastic dynamics of adhesion clusters as a function of cluster size, rebinding rate and force has been solved both analytically and numerically. Moreover this model has been applied to the regulation of cell-matrix contacts, to dynamic force spectroscopy, and to rolling adhesion. (3) Using linear elasticity theory and the concept of force dipoles, a model has been introduced and solved which predicts the positioning and orientation of mechanically active cells in soft material, in good agreement with experimental observations for fibroblasts on elastic substrates and in collagen gels.
Concerns have been raised that anthropogenic climate change could lead to large-scale singular climate events, i.e., abrupt nonlinear climate changes with repercussions on regional to global scales. One central goal of this thesis is the development of models of two representative components of the climate system that could exhibit singular behavior: the Atlantic thermohaline circulation (THC) and the Indian monsoon. These models are conceived so as to fulfill the main requirements of integrated assessment modeling, i.e., reliability, computational efficiency, transparency and flexibility. The model of the THC is an interhemispheric four-box model calibrated against data generated with a coupled climate model of intermediate complexity. It is designed to be driven by global mean temperature change which is translated into regional fluxes of heat and freshwater through a linear down-scaling procedure. Results of a large number of transient climate change simulations indicate that the reduced-form THC model is able to emulate key features of the behavior of comprehensive climate models such as the sensitivity of the THC to the amount, regional distribution and rate of change in the heat and freshwater fluxes. The Indian monsoon is described by a novel one-dimensional box model of the tropical atmosphere. It includes representations of the radiative and surface fluxes, the hydrological cycle and surface hydrology. Despite its high degree of idealization, the model satisfactorily captures relevant aspects of the observed monsoon dynamics, such as the annual course of precipitation and the onset and withdrawal of the summer monsoon. Also, the model exhibits the sensitivity to changes in greenhouse gas and sulfate aerosol concentrations that are known from comprehensive models. A simplified version of the monsoon model is employed for the identification of changes in the qualitative system behavior against changes in boundary conditions. The most notable result is that under summer conditions a saddle-node bifurcation occurs at critical values of the planetary albedo or insolation. Furthermore, the system exhibits two stable equilibria: besides the wet summer monsoon, a stable state exists which is characterized by a weak hydrological cycle. These results are remarkable insofar, as they indicate that anthropogenic perturbations of the planetary albedo such as sulfur emissions and/or land-use changes could destabilize the Indian summer monsoon. The reduced-form THC model is employed in an exemplary integrated assessment application. Drawing on the conceptual and methodological framework of the tolerable windows approach, emissions corridors (i.e., admissible ranges of CO2- emissions) are derived that limit the risk of a THC collapse while considering expectations about the socio-economically acceptable pace of emissions reductions. Results indicate, for example, a large dependency of the width of the emissions corridor on climate and hydrological sensitivity: for low values of climate and/or hydrological sensitivity, the corridor boundaries are far from being transgressed by any plausible emissions scenario for the 21st century. In contrast, for high values of both quantities low non-intervention scenarios leave the corridor already in the early decades of the 21st century. This implies that if the risk of a THC collapse is to be kept low, business-as-usual paths would need to be abandoned within the next two decades. All in all, this thesis highlights the value of reduced-form modeling by presenting a number of applications of this class of models, ranging from sensitivity and bifurcation analysis to integrated assessment. The results achieved and conclusions drawn provide a useful contribution to the scientific and policy debate about the consequences of anthropogenic climate change and the long-term goals of climate protection. --- Anmerkung: Die Autorin ist Trägerin des von der Mathematisch-Naturwissenschaftlichen Fakultät der Universität Potsdam vergebenen Michelson-Preises für die beste Promotion des Jahres 2003/2004.
A polymer is a large molecule made up of many elementary chemical units, joined together by covalent bonds (for example, polyethylene). Polyelectrolytes (PELs) are polymer chains containing a certain amount of ionizable monomers. With their specific properties PELs acquire big importance in molecular and cell biology as well as in technology. Compared to neutral polymers the theory of PELs is less understood. In particular, this is valid for PELs in poor solvents. A poor solvent environment causes an effective attraction between monomers. Hence, for PELs in a poor solvent, there occurs a competition between attraction and repulsion. Strong or quenched PELs are completely dissociated at any accessible pH. The position of charges along the chain is fixed by chemical synthesis. On the other hand, in weak or annealed PELs dissociation of charges depends on solution pH. For the first time the simulation results have given direct evidence that at rather poor solvents an annealed PEL indeed undergoes a first-order phase transition when the chemical potential (solution pH) reaches at a certain value. The discontinuous transition occurs between a weakly charged compact globular structure and a strongly charged stretched configuration. At not too poor solvents theory predicts that globule would become unstable with respect to the formation of pearl-necklaces. The results show that pearl-necklaces exist in annealed PELs indeed. Furthermore, as predicted by theory, the simulation results have shown that annealed PELs display a sharp transition from a highly charged stretched state to a weakly charged globule at a critical salt concentration.
Die vorliegende Arbeit beschäftigt sich mit der Charakterisierung von Seismizität anhand von Erdbebenkatalogen. Es werden neue Verfahren der Datenanalyse entwickelt, die Aufschluss darüber geben sollen, ob der seismischen Dynamik ein stochastischer oder ein deterministischer Prozess zugrunde liegt und was daraus für die Vorhersagbarkeit starker Erdbeben folgt. Es wird gezeigt, dass seismisch aktive Regionen häufig durch nichtlinearen Determinismus gekennzeichent sind. Dies schließt zumindest die Möglichkeit einer Kurzzeitvorhersage ein. Das Auftreten seismischer Ruhe wird häufig als Vorläuferphaenomen für starke Erdbeben gedeutet. Es wird eine neue Methode präsentiert, die eine systematische raumzeitliche Kartierung seismischer Ruhephasen ermöglicht. Die statistische Signifikanz wird mit Hilfe des Konzeptes der Ersatzdaten bestimmt. Als Resultat erhält man deutliche Korrelationen zwischen seismischen Ruheperioden und starken Erdbeben. Gleichwohl ist die Signifikanz dafür nicht hoch genug, um eine Vorhersage im Sinne einer Aussage über den Ort, die Zeit und die Stärke eines zu erwartenden Hauptbebens zu ermöglichen.
Diese Arbeit beschäftigt sich mit der Annahme, dass den Erdbeben ein selbstorganisiert kritischer Zustand der Erdkruste zugrunde liegt. Mit Hilfe einer Erweiterung bisheriger Modelle wird gezeigt, dass ein solcher Zustand nicht nur für die Grössenverteilung der Erdbeben (Gutenberg-Richter Gesetz), sondern auch für das beobachtete raumzeitliche Auftreten, z.B. für das Omori-Gesetz für Nachbebenserien, verantwortlich sein kann. Desweiteren wird die Frage nach der Vorhersagbarkeit grosser Erdbeben in solchen Modellsimulationen untersucht.
We theoretically discuss the interaction of neutral particles (atoms, molecules) with surfaces in the regime where it is mediated by the electromagnetic field. A thorough characterization of the field at sub-wavelength distances is worked out, including energy density spectra and coherence functions. The results are applied to typical situations in integrated atom optics, where ultracold atoms are coupled to a thermal surface, and to single molecule probes in near field optics, where sub-wavelength resolution can be achieved.
This thesis presents new approaches to evolutions of binary black hole systems in numerical relativity. We analyze and compare evolutions from various physically motivated initial data sets, in particular presenting the first evolutions of Thin Sandwich data generated by the Meudon group. For the first time two different quasi-circular orbit initial data sequences are compared through fully 3d numerical evolutions: Puncture data and Thin Sandwich data (TSD) based on a helical killing vector ansatz. The two different sets are compared in terms of the physical quantities that can be measured from the numerical data, and in terms of their evolutionary behavior. The evolutions demonstrate that for the latter, "Meudon" datasets, the black holes do in fact orbit for a longer amount of time before they merge, in comparison with Puncture data from the same separation. This indicates they are potentially better estimates of quasi-circular orbit parameters. The merger times resulting from the numerical simulations are consistent with independent Post-Newtonian estimates that the final plunge phase of a black hole inspiral should take 60% of an orbit.
In-situ Wachstumsuntersuchungen beim reaktiven Anlassen von Cu, In Schichten in elementarem Schwefel
(2003)
In dieser Arbeit wurde das reaktive Anlassen von dünnen Kupfer-Indium-Schichten in elementarem Schwefel mittels energiedispersiver Röntgenbeugung untersucht. Durch die simultane Aufnahme der Röntgenspektren und der Messung der diffusen Reflexion von Laserlicht der Wellenlänge 635 nm an der Oberfläche der Probe während des Schichtwachstums von CuInS<SUB>2</SUB> konnte eine Methode zur Prozesskontrolle für ein Herstellungsverfahren von CuInS<SUB>2</SUB> etabliert werden. Die Bildung von CuInS<SUB>2</SUB> aus Kupfer-Indium-Vorläuferschichten wird dominiert von Umwandlungen der intermetallischen Phasen. CuInS<SUB>2</SUB> wächst innerhalb der Aufheizperiode ab einer Temperatur von ca. 200°C aus der Phase Cu11In9. Jedoch zerfällt letztere metallische Phase in Cu16In9 und flüssiges Indium bei einer Temperatur von ca. 310°C. Das flüssige Indium reagiert im Falle von Kupferarmut mit dem Schwefel und führt zu einem zusätzlichen Reaktionspfad über InS zu CuIn5S8. Unter Präparationsbedingungen mit Kupferüberschuss wird das Indium in einer intermetallischen Phase gebunden.Erstmals konnte die Phase Digenite bei Temperaturen über 240°C beobachtet werden. Beim Abkühlen auf Raumtemperatur wandelt sich diese Phase unter dem Verbrauch von Schwefel in Covellite um.Für Proben mit Kupferüberschuss konnte eine Wachstumskinetik proportional zur Temperatur beobachtet werden. Dieses Verhalten wurde durch eine stress-induzierte Diffusion als dominierenden Reaktionsmechanismus interpretiert. Dabei werden während der Bildung von CuInS<SUB>2</SUB> durch unterschiedliche Ausdehnungen der metallischen und sulfidischen Schichten eine Spannung in der CuInS<SUB>2</SUB>-Schicht induziert, die nach Überschreiten einer Grenzspannung zu Rissen in der CuInS2-Schicht führt. Entlang dieser Risse findet ein schneller Transport der Metalle zur Oberfläche, wo diese mit dem Schwefel reagieren können, statt. Die Risse heilen durch die Bildung neuen Sulfids wieder aus.
Die heutige optische Informationsverarbeitung erfordert neue Materialien, die Licht effektiv verarbeiten, steuern und speichern können. Photorefraktive (PR) Materialien sind dafür sehr interessant. In diesen Materialien entsteht bei inhomogener Beleuchtung (z.B. mit einem Intererenzmuster) über Ladungsträgergenerierung und Einfang der Ladungsträger in Fallen ein Raumladungsfeld. Dieses wird über den elektrooptischen Effekt in eine räumliche Modulation des Brechungsindex umgesetzt. Letztendlich führt somit die inhomogene Beleuchtung eines PR-Materials zu einer räumlich variierenden Änderung des Brechungsindex. Vor ca. 10 Jahren wurde entdeckt, dass auch Polymere einen PR-Effekt aufweisen können. Die Ansprechzeit dieser Materialien wird dabei wesentlich durch die Dynamik der Ladungsträger (bestimmt durch Erzeugung, Transport, Einfang in Fallen etc.) begrenzt. Bis zu Beginn dieser Arbeit war es noch nicht gelungen, einen quantitativen Zusammenhang zwischen der Ladungsträgerdynamik und der Ansprechzeit des PR-Effekts experimentell nachzuweisen. In dieser Arbeit wird ein Weg aufgezeigt, durch photophysikalische Experimente unter verschiedenen Beleuchtungsbedingungen alle photophysikalischen Größen experimentell zu bestimmen, die den Aufbau des Raumladungsfelds in organischen photorefraktiven Materialien bestimmen. So konnte durch Experimente unter Beleuchtung mit kurzen Einzelpulsen sowohl die Beweglichkeit der freien Ladungsträger als auch die charakteristischen Parameter flacher Fallen ermittelt werden. Zur Bestimmung der Dichte tiefer Fallen wurde die Intensitätsabhängigkeit des stationären Photostroms untersucht. Durch die analytische Lösung des bestimmenden Gleichungssystems konnte gezeigt werden, dass die Sublinearität der Intensitätsabhängigkeit des Photostroms primär mit dem Verhältnis zwischen Entleerungs- und Einfangkoeffizienten tiefer Fallen korreliert. Zur unabhängigen Bestimmung des Entleerungskoeffizienten der tiefen Fallen wurden Doppelpulsexperimente mit variabler Verzögerungszeit zwischen den Pulsen verwendet. Mit den erhaltenen Parametern konnte dann das untere Limit der zum Aufbau des Raumladungsfelds notwendigen Zeit abgeschätzt werden. Diese Werte wurden mit den gemessenen photorefraktiven Ansprechzeiten verglichen. Es zeigt sich, dass weder die Photogeneration noch der Transport der Ladungsträger die Geschwindigkeit des Aufbaus des Raumladungsfeldes limitiert. Stattdessen konnte erstmals quantitativ nachgewiesen werden, dass die Dynamik des Raumladungsfelds in den hier untersuchten PR-Materialien durch das Füllen tiefer Fallen mit photogenerierten Ladungsträgern bestimmt wird. Dabei spielt das Verhältnis zwischen dem Einfang- und dem Rekombinationskoeffizienten eine wesentliche Rolle. Weiterhin wurde die Dynamik des Aufbaus des Raumladungsfelds bei unterschiedlichen Vorbeleuchtungsbedingungen quantitativ simuliert und mit den experimentellen PR-Transienten verglichen. Die gute Übereinstimmung zwischen den simulierten und gemessenen Transienten erlaubte es abschließend, die kritischen Parameter, die die Dynamik des PR-Effekts in den untersuchten Polymeren begrenzen, zu identifizieren.
This work incorporates three treatises which are commonly concerned with a stochastic theory of the Lyapunov exponents. With the help of this theory universal scaling laws are investigated which appear in coupled chaotic and disordered systems. First, two continuous-time stochastic models for weakly coupled chaotic systems are introduced to study the scaling of the Lyapunov exponents with the coupling strength (coupling sensitivity of chaos). By means of the the Fokker-Planck formalism scaling relations are derived, which are confirmed by results of numerical simulations. Next, coupling sensitivity is shown to exist for coupled disordered chains, where it appears as a singular increase of the localization length. Numerical findings for coupled Anderson models are confirmed by analytic results for coupled continuous-space Schrödinger equations. The resulting scaling relation of the localization length resembles the scaling of the Lyapunov exponent of coupled chaotic systems. Finally, the statistics of the exponential growth rate of the linear oscillator with parametric noise are studied. It is shown that the distribution of the finite-time Lyapunov exponent deviates from a Gaussian one. By means of the generalized Lyapunov exponents the parameter range is determined where the non-Gaussian part of the distribution is significant and multiscaling becomes essential.
Understanding the formation of stars in galaxies is central to much of modern astrophysics. For several decades it has been thought that the star formation process is primarily controlled by the interplay between gravity and magnetostatic support, modulated by neutral-ion drift. Recently, however, both observational and numerical work has begun to suggest that supersonic interstellar turbulence rather than magnetic fields controls star formation. This review begins with a historical overview of the successes and problems of both the classical dynamical theory of star formation, and the standard theory of magnetostatic support from both observational and theoretical perspectives. We then present the outline of a new paradigm of star formation based on the interplay between supersonic turbulence and self-gravity. Supersonic turbulence can provide support against gravitational collapse on global scales, while at the same time it produces localized density enhancements that allow for collapse on small scales. The efficiency and timescale of stellar birth in Galactic gas clouds strongly depend on the properties of the interstellar turbulent velocity field, with slow, inefficient, isolated star formation being a hallmark of turbulent support, and fast, efficient, clustered star formation occurring in its absence. After discussing in detail various theoretical aspects of supersonic turbulence in compressible self-gravitating gaseous media relevant for star forming interstellar clouds, we explore the consequences of the new theory for both local star formation and galactic scale star formation. The theory predicts that individual star-forming cores are likely not quasi-static objects, but dynamically evolving. Accretion onto these objects will vary with time and depend on the properties of the surrounding turbulent flow. This has important consequences for the resulting stellar mass function. Star formation on scales of galaxies as a whole is expected to be controlled by the balance between gravity and turbulence, just like star formation on scales of individual interstellar gas clouds, but may be modulated by additional effects like cooling and differential rotation. The dominant mechanism for driving interstellar turbulence in star-forming regions of galactic disks appears to be supernovae explosions. In the outer disk of our Milky Way or in low-surface brightness galaxies the coupling of rotation to the gas through magnetic fields or gravity may become important.
This thesis describes the development and application of the impacts module of the ICLIPS model, a global integrated assessment model of climate change. The presentation of the technical aspects of this model component is preceded by a discussion of the sociopolitical context for model-based integrated assessments, which defines important requirements for the specification of the model. Integrated assessment of climate change comprises a broad range of scientific efforts to support the decision-making about objectives and measures for climate policy, whereby many different approaches have been followed to provide policy-relevant information about climate impacts. Major challenges in this context are the large diversity of the relevant spatial and temporal scales, the multifactorial causation of many climate impacts', considerable scientific uncertainties, and the ambiguity associated with unavoidable normative evaluations. A hierarchical framework is presented for structuring climate impact assessments that reflects the evolution of their practice and of the underlying theory. Integrated assessment models of climate change (IAMs) are scientific tools that contain simplified representations of the relevant components of the coupled society-climate system. The major decision-analytical frameworks for IAMs are evaluated according to their ability to address important aspects of the pertinent social decision problem. The guardrail approach is presented as an inverse' framework for climate change decision support, which aims to identify the whole set of policy strategies that are compatible with a set of normatively specified constraints (guardrails'). This approach combines, to a certain degree, the scientific rigour and objectivity typical of predictive approaches with the ability to consider virtually all decision options that is at the core of optimization approaches. The ICLIPS model is described as the first IAM that implements the guardrail approach. The representation of climate impacts is a key concern in any IAM. A review of existing IAMs reveals large differences in the coverage of impact sectors, in the choice of the impact numeraire(s), in the consideration of non-climatic developments, including purposeful adaptation, in the handling of uncertainty, and in the inclusion of singular events. IAMs based on an inverse approach impose specific requirements to the representation of climate impacts. This representation needs to combine a level of detail and reliability that is sufficient for the specification of impact guardrails with the conciseness and efficiency that allows for an exploration of the complete domain of plausible climate protection strategies. Large-scale singular events can often be represented by dynamic reduced-form models. This approach, however, is less appropriate for regular impacts where the determination of policy-relevant results generally needs to consider the heterogeneity of climatic, environmental, and socioeconomic factors at the local or regional scale. Climate impact response functions (CIRFs) are identified as the most suitable reduced-form representation of regular climate impacts in the ICLIPS model. A CIRF depicts the aggregated response of a climate-sensitive system or sector as simulated by a spatially explicit sectoral impact model for a representative subset of plausible futures. In the CIRFs presented here, global mean temperature and atmospheric CO2 concentration are used as predictors for global and regional impacts on natural vegetation, agricultural crop production, and water availability. Application of a pattern scaling technique makes it possible to consider the regional and seasonal patterns in the climate anomalies simulated by several general circulation models while ensuring the efficiency of the dynamic model components. Efforts to provide quantitative estimates of future climate impacts generally face a trade-off between the relevance of an indicator for stakeholders and the exactness with which it can be determined. A number of non-monetary aggregated impact indicators for the CIRFs is presented, which aim to strike the balance between these two conflicting goals while taking into account additional constraints of the ICLIPS modelling framework. Various types of impact diagrams are used for the visualization of CIRFs, each of which provides a different perspective on the impact result space. The sheer number of CIRFs computed for the ICLIPS model precludes their comprehensive presentation in this thesis. Selected results referring to changes in the distribution of biomes in different biogeographical regions, in the agricultural potential of various countries, and in the water availability in selected major catchments are discussed. The full set of CIRFs is accessible via the ICLIPS Impacts Tool, a graphical user interface that provides convenient access to more than 100,000 impact diagrams developed for the ICLIPS model. The technical aspects of the software are described as well as the accompanying database of CIRFs. The most important application of CIRFs is in inverse' mode, where they are used to translate impact guardrails into simultaneous constraints for variables from the optimizing ICLIPS climate-economy model. This translation is facilitated by algorithms for the computation of reachable climate domains and for the parameterized approximation of admissible climate windows derived from CIRFs. The comprehensive set of CIRFs, together with these algorithms, enables the ICLIPS model to flexibly explore sets of climate policy strategies that explicitly comply with impact guardrails specified in biophysical units. This feature is not found in any other intertemporally optimizing IAM. A guardrail analysis with the integrated ICLIPS model is described that applies selected CIRFs for ecosystem changes. So-called necessary carbon emission corridors' are determined for a default choice of normative constraints that limit global vegetation impacts as well as regional mitigation costs, and for systematic variations of these constraints. A brief discussion of recent developments in integrated assessment modelling of climate change connects the work presented here with related efforts.
In this thesis the gravitational lensing effect is used to explore a number of cosmological topics. We determine the time delay in the gravitationally lensed quasar system HE1104-1805 using different techniques. We obtain a time delay Delta_t(A-B) Delta_t(A-B) =-310 +- 20 days (2 sigma errors) between the two components. We also study the double quasar Q0957+561 during a three years monitoring campaign. The fluctuations we find in the difference light curves are completely consistent with noise and no microlensing is needed to explain these fluctuations. Microlensing is also studied in the quadruple quasar Q2237+0305 during the GLITP collaboration (Oct.1999-Feb.2000). We use the absence of a strong microlensing signal to obtain an upper limit of v=600 km/s for the effective transverse velocity of the lens galaxy (considering microlenses with 0.1 solar masses). The distribution of dark matter in galaxy clusters is also studied in the second part of the thesis. In the cluster of galaxies Cl0024+1654 we obtain a mass-to-light ratio of M/L = 200 M_sun/L_sun (within a radius of 3 arcminutes). In the galaxy cluster RBS380 we find a relatively low X-ray luminosity for a massive cluster of L =2*10^(44) erg/s, but a rich distribution of galaxies in the optical band.
In festen azobenzenhaltigen Polymeren wurde bei Bestrahlung mit blauem Licht ein makroskopischer Materialtransport beobachtet. Um die Dynamik der Gitterentstehung zu verfolgen, wurde am Speicherring für Synchrotronstrahlung ein Gitterschreibaufbau errichtet. Damit konnte erstmals in dieser Arbeit die Gitterbildungsgeschwindigkeit in-situ simultan mit Röntgen- und Lichtstreuung untersucht werden. Mit Hilfe einer speziellen Anpassung der Röntgenstreutheorie konnten sehr gute Übereinstimmungen von theoretischen Berechnungen mit den Messergebnissen erzielt werden. Dabei konnte nachgewiesen werden, dass sich zeitgleich mit einem Oberflächengitter auch ein Dichtegitter entwickelt. Durch die Trennung beider Streuanteile ließ sich die Dynamik der Strukturentstehungen bestimmen. Des weiteren konnte erstmals mit Hilfe der Photoelektronenspektroskopie die molekulare Orientierung an der Oberfläche eines Oberflächengitters nachgewiesen werden. Die Bewegungsursache kann auf einen Impulsübertrag während der Isomerisierung zurückgeführt werden, während die Bewegungsrichtung durch den elektrischen Feldvektor festgelegt wird. Die Theorie der Gitterentstehung konnte verbessert werden.
Eine Nutzung der optischen Anisotropie dünner Schichten ist vor allem für die Displaytechnologie, die optische Datenspeicherung und für optische Sicherheitselemente von hoher Bedeutung. Diese Doktorarbeit befasst sich mit theoretischen und experimentellen Untersuchung von dreidimensionaler Anisotropie und dabei insbesondere mit der Untersuchung von lichtinduzierter dreidimensionaler Anisotropie in organischen dünnen Polymer-Schichten. Die gewonnenen Erkentnisse und entwickelten Methoden können wertvolle Beiträge für Optimierungsprozesse, wie bei der Kompensation der Blickwinkelabhängigkeit von Flüssigkristall-Displays, liefern. Die neue Methode der Immersions-Transmissions-Ellipsometrie (ITE) zur Untersuchung von dünneren Schichten wurde im Rahmen dieser Dissertation entwickelt. Diese Methode gestattet es, in Kombination mit konventioneller Reflexions- und Transmissionsellipsometrie, die absoluten dreidimensionalen Brechungsindices einer biaxialen Schicht zu bestimmen. Erstmals gelang es damit, das dreidimensionale Brechungsindexellipsoid von transparenten, dünneren (150 nm) Filmen hochgenau (drei Stellen hinter dem Komma) zu bestimmen. Die ITE-Methode hat demzufolge das Potential, auch bei noch dünneren Schichten mit Gewinn eingesetzt werden zu können. Die lichtinduzierte Generierung von dreidimensionaler Anisotropie wurde in dünnen Schichten von azobenzenhaltigen und zimtsäurehaltigen, amorphen und flüssig-kristallinen Homo- und Copolymeren untersucht. Erstmals wurden quantitative Untersuchungen zur Änderung von lichtinduzierten, dreidimensionalen Anisotropien in dünnen Schichten von azobenzenhaltigen und zimtsäurehaltigen Polymeren bei Tempern oberhalb der Glastemperatur durchgeführt. Bei vielen der untersuchten Polymere war die dreidimensionale Ordnung nach dem Bestrahlen mit polarisiertem Licht und anschließendem Tempern oberhalb der Glastemperatur scheinbar von der Schichtdicke abhängig. Die Ursache liegt wohl in der, mit der neuentwickelten ITE-Methode detektierten, planaren Ausgangsorientierung der aufgeschleuderten dünneren Schichten. Um Verkippungs-Gradienten in dickeren Polymerschichten in ihrem Verlauf zu bestimmen, wurde eine spezielle Methode unter Benutzung der Wellenleitermoden-Spektroskopie entwickelt. Quantenchemisch bestimmte, maximal induzierbare Doppelbrechungen in flüssig-kristallinen Polymeren wurden mit den experimentell gefundenen Ordnungen verglichen.
The correlations between the chemical structures of the 2,5-diphenyl-1,3,4-oxadiazole compounds and their corresponding vapour deposited film structures on Si/SiO2 were systematically investigated with AFM, XSR and IR for the first time. The result shows that the film structure depends strongly on the substrate temperature (Ts). For the compounds with ether bridge group, the film periodicity depends linearly on the length of the aliphatic chain. The films based on those oxadiazols have ordered structure in the investigated substrate temperature region, while die amide bridged compounds form ordered film only at high Ts due to the formation of intermolecular H-bond. The tilt angle of most molecules is determined by the pi-pi complexes between the molecules. The intermolecular interaction between head groups leads to the structural transformation during the thermal treatment after deposition. All the ether bridged oxadiazoles form films with bilayer structure, while amide bridged oxadiazole form film bilayer structure only when the molecule has a head group.
This thesis deals with the encoding and transmission of information through a quantum channel. A quantum channel is a quantum mechanical system whose state is manipulated by a sender and read out by a receiver. The individual state of the channel represents the message. The two topics of the thesis comprise 1) the possibility of compressing a message stored in a quantum channel without loss of information and 2) the possibility to communicate a message directly from one party to another in a secure manner, that is, a third party is not able to eavesdrop the message without being detected. The main results of the thesis are the following. A general framework for variable-length quantum codes is worked out. These codes are necessary to make lossless compression possible. Due to the quantum nature of the channel, the encoded messages are in general in a superposition of different lengths. It is found to be impossible to compress a quantum message without loss of information if the message is not apriori known to the sender. In the other case it is shown that lossless quantum data compression is possible and a lower bound on the compression rate is derived. Furthermore, an explicit compression scheme is constructed that works for arbitrarily given source message ensembles. A quantum cryptographic protocol - the “ping-pong protocol” - is presented that realizes the secure direct communication of classical messages through a quantum channel. The security of the protocol against arbitrary eavesdropping attacks is proven for the case of an ideal quantum channel. In contrast to other quantum cryptographic protocols, the ping-pong protocol is deterministic and can thus be used to transmit a random key as well as a composed message. The protocol is perfectly secure for the transmission of a key, and it is quasi-secure for the direct transmission of a message. The latter means that the probability of successful eavesdropping exponentially decreases with the length of the message.
We study the effect on the elastic properties of lipid membranes induced by anchoring of long hydrophilic polymers. Theoretically, two limiting regimes for the membrane spontaneous curvature are expected : i) at low surface polymer concentration (mushroom regime) the spontaneous curvature should scale linearly with the surface density of anchored polymers; ii) at high coverage (brush regime) the dependence should be quadratic. We attempt to test the predictions for the brush regime by monitoring the morphological changes induced on giant vesicles. As long polymers we use fluorescently labeled λ-phage DNA molecules which are attached to biotinylated lipid vesicles with a biotin-avidin-biotin linkage. By varying the amount of biotinylated lipid in the membrane we control the surface concentration of the anchors. The amount of anchored DNA to the membrane is quantified with fluorescence measurements. Changes in the elastic properties of the membrane as DNA grafts to it are monitored via analysis of the vesicle fluctuations. The spontaneous curvature of the membrane increases as a function of the surface coverage. At higher grafting concentrations the vesicles bud. The size of the buds can also be used to assess the membrane curvature. The effect on the bending stiffness is a subject of further investigation.
The theory of atomic Boson-Fermion mixtures in the dilute limit beyond mean-field is considered in this thesis. Extending the formalism of quantum field theory we derived expressions for the quasi-particle excitation spectra, the ground state energy, and related quantities for a homogenous system to first order in the dilute gas parameter. In the framework of density functional theory we could carry over the previous results to inhomogeneous systems. We then determined to density distributions for various parameter values and identified three different phase regions: (i) a stable mixed regime, (ii) a phase separated regime, and (iii) a collapsed regime. We found a significant contribution of exchange-correlation effects in the latter case. Next, we determined the shift of the Bose-Einstein condensation temperature caused by Boson-Fermion interactions in a harmonic trap due to redistribution of the density profiles. We then considered Boson-Fermion mixtures in optical lattices. We calculated the criterion for stability against phase separation, identified the Mott-insulating and superfluid regimes both, analytically within a mean-field calculation, and numerically by virtue of a Gutzwiller Ansatz. We also found new frustrated ground states in the limit of very strong lattices. ----Anmerkung: Der Autor ist Träger des durch die Physikalische Gesellschaft zu Berlin vergebenen Carl-Ramsauer-Preises 2004 für die jeweils beste Dissertation der vier Universitäten Freie Universität Berlin, Humboldt-Universität zu Berlin, Technische Universität Berlin und Universität Potsdam.