Refine
Year of publication
Document Type
- Article (21297)
- Doctoral Thesis (3184)
- Postprint (2347)
- Monograph/Edited Volume (1223)
- Other (675)
- Review (631)
- Preprint (531)
- Conference Proceeding (474)
- Part of a Book (255)
- Working Paper (182)
Language
- English (31017) (remove)
Keywords
- climate change (177)
- Germany (105)
- machine learning (86)
- diffusion (76)
- German (68)
- morphology (67)
- Arabidopsis thaliana (66)
- anomalous diffusion (58)
- stars: massive (58)
- Climate change (55)
Institute
- Institut für Physik und Astronomie (4993)
- Institut für Biochemie und Biologie (4731)
- Institut für Geowissenschaften (3333)
- Institut für Chemie (2917)
- Institut für Mathematik (1876)
- Department Psychologie (1490)
- Institut für Ernährungswissenschaft (1048)
- Department Linguistik (1013)
- Wirtschaftswissenschaften (858)
- Institut für Informatik und Computational Science (841)
Late Miocene to Quaternary volcanic rocks from the frontal arc to the back-arc region of the Central Volcanic Zone in the Andes show a wide range of delta 11B values (+4 to -7 ‰) and boron concentrations (6 to 60 ppm). Positive delta 11B values of samples from the volcanic front indicate involvement of a 11B-enriched slab component, most likely derived from altered oceanic crust, despite the thick Andean continental lithosphere, and rule out a pure crust-mantle origin for these lavas. The delta 11B values and B concentrations in the lavas decrease systematically with increasing depth of the Wadati-Benioff Zone. This across-arc variation in delta 11B values and decreasing B/Nb ratios from the arc to the back-arc samples are attributed to the combined effects of B-isotope fractionation during progressive dehydration in the slab and a steady decrease in slab-fluid flux towards the back arc, coupled with a relatively constant degree of crustal contamination as indicated by similar Sr, Nd and Pb isotope ratios in all samples. Modelling of fluid-mineral B-isotope fractionation as a function of temperature fits the across-arc variation in delta 11B and we conclude that the B-isotope composition of arc volcanics is dominated by changing delta 11B composition of B-rich slab-fluids during progressive dehydration. Crustal contamination becomes more important towards the back-arc due to the decrease in slab-derived fluid flux. Because of this isotope fractionation effect, high delta 11B signatures in volcanic arcs need not necessarily reflect differences in the initial composition of the subducting slab. Three-component mixing calculations for slab-derived fluid, the mantle wedge and the continental crust based on B, Sr and Nd isotope data indicate that the slab-fluid component dominates the B composition of the fertile mantle and that the primary arc magmas were contaminated by an average addition of 15 to 30 % crustal material.
Transport processes in and of cells are of major importance for the survival of the organism. Muscles have to be able to contract, chromosomes have to be moved to opposing ends of the cell during mitosis, and organelles, which are compartments enclosed by membranes, have to be transported along molecular tracks. Molecular motors are proteins whose main task is moving other molecules.For that purpose they transform the chemical energy released in the hydrolysis of ATP into mechanical work. The motors of the cytoskeleton belong to the three super families myosin, kinesin and dynein. Their tracks are filaments of the cytoskeleton, namely actin and the microtubuli. Here, we examine stochastic models which are used for describing the movements of these linear molecular motors. The scale of the movements comprises the regime of single steps of a motor protein up to the directed walk along a filament. A single step bridges around 10 nm, depending on the protein, and takes about 10 ms, if there is enough ATP available. Our models comprise M states or conformations the motor can attain during its movement along a one-dimensional track. At K locations along the track transitions between the states are possible. The velocity of the protein depending on the transition rates between the single states can be determined analytically. We calculate this velocity for systems of up to four states and locations and are able to derive a number of rules which are helpful in estimating the behaviour of an arbitrary given system. Beyond that we have a look at decoupled subsystems, i.e., one or a couple of states which have no connection to the remaining system. With a certain probability a motor undergoes a cycle of conformational changes, with another probability an independent other cycle. Active elements in real transport processes by molecular motors will not be limited to the transitions between the states. In distorted networks or starting from the discrete Master equation of the system, it is possible to specify horizontal rates, too, which furthermore no longer have to fulfill the conditions of detailed balance. Doing so, we obtain unique, complete paths through the respective network and rules for the dependence of the total current on all the rates of the system. Besides, we view the time evolutions for given initial distributions. In enzymatic reactions there is the idea of a main pathway these reactions follow preferably. We determine optimal paths and the maximal flow for given networks. In order to specify the dependence of the motor's velocity on its fuel ATP, we have a look at possible reaction kinetics determining the connection between unbalanced transitions rates and ATP-concentration. Depending on the type of reaction kinetics and the number of unbalanced rates, we obtain qualitatively different curves connecting the velocity to the ATP-concentration. The molecular interaction potentials the motor experiences on its way along its track are unknown. We compare different simple potentials and the effects the localization of the vertical rates in the network model has on the transport coefficients in comparison to other models.
The P- and S-wave velocity structure of the D” layer beneath the southwestern Pacific was investigated by using short-period data from 12 deep events in the Tonga-Fiji region recorded by the J-Array and the Hi-net in Japan. A migration method and reflected wave beamforming (RWB) were used in order to extract weak signals originating from small-scale heterogeneities in the lowermost mantle. In order to acquire high resolution, a double array method (DAM) which integrates source array beamforming with receiver array beamforming was applied to the data. A phase-weighted stacking technique, which reduces incoherent noise by employing complex trace analysis, was also applied to the data, amplifying the weak coherent signals from the lowermost mantle. This combination greatly enhances small phases common to the source and receiver beams. The results of the RWB method indicate that seismic energy is reflected at discontinuities near 2520 km and 2650 km, which have a negative P-wave velocity contrast of 1 % at the most. In addition, there is a positive seismic discontinuity at a depth of 2800 km. In the case of the S-wave, reflected energy is produced almost at the same depth (2550 km depth). The different depth (50 km) between the P-wave velocity discontinuity at the depth of 2800 and a further S-wave velocity discontinuity at the depth of 2850 km may indicate that the S-wave velocity reduction in the lowermost mantle is about 2-3 times stronger that that of P wave. A look at a 2D cross section, constructed with the RWB method, suggests that the observed discontinuities can be characterized as intermittent lateral heterogeneities whose lateral extent is a few hundred km, and that the CMB might have undulations on a scale of less than 10 km in amplitude. The migration shows only weak evidence for the existence of scattering objects. Heterogeneous regions in the migration belong to the detected seismic discontinuities. These anomalous structures may represent a part of hot plume generated beneath the southwestern Pacific in the lowermost mantle.
This study investigated the slope carbonates of two Miocene carbonate systems: the Maltese Islands (in the Central Mediterranean) and the Marion Plateau (Northeastern Australia, drilled during ODP Leg 194). The aim of the study was to trace the impact of the Miocene cooling steps (events Mi1-Mi6) in these carbonate systems, especially the Mi3 event, which took place around 13.6 Ma and deeply impacted the marine oxygen isotope record. This event also profoundly impacted oceanographic and climatic patterns, eventually leading to the establishment of the modern ice-house world. In particular, East Antarctica became ice covered at that period. The rational behind the present study was to investigate the impact that this event had on shallow water systems in order to complement the deep-sea record and hence acquire a more global perspective on Miocene climate change. The Maltese Islands were investigated for trends in bulk-rock carbon and oxygen isotopes, as well as bulk-rock mineralogy, clay minerals analysis and organic geochemisty. Results showed that the mid Miocene cooling event deeply impacted sedimentation at that location by changing sedimentation from carbonate to clay-rich sediments. Moreover, it was discovered that each phase of Antarctic glaciation, not just the major mid Miocene event, resulted in higher terrigenous input on Malta. Mass accumulation rates revealed that this was linked to increased runoff during periods when Antarctica was glaciated, and thus that the carbonate sediments were “diluted” by clay-rich sediments. The model subsequently developed to explain this implies feedback from Antarctic glaciations creating cold, dense air masses that push the ITCZ Northward, thus increasing precipitation on the North African subcontinent. Increased precipitation (or stronger African monsoon) accelerated continental weathering and runoff, thus bringing more terrigenous sediment to the paleo-location of the slope sediments of Malta. Spectral analysis of carbonate content and organic matter geochemical analysis furthermore suggest that the clay-rich intervals are similar to sapropelic deposits. On the Marion Plateau, trends in oxygen and carbon isotopes were obtained by measuring Cibicidoides spp foraminifers. Moreover, carbonate content was reconstructed using a chemical method (coulometer). Results show that the mid Miocene cooling step profoundly affected this system: a major drop in accumulation rates of carbonates occurs precisely at 13.8 Ma, around the time of the East Antarctic ice sheet formation. Moreover, sedimentation changes occurred at that time, carbonate fragments coming from neritic environments becoming less abundant, planktonic foraminifer content increasing and quartz and reworked glauconite being deposited. Conversely, a surprising result is that the major N12-N14 sea-level fall occurring around 11.5 Ma did not impact the accumulation of carbonates on the slope. This was unexpected since carbonate platform are very sensitive to sea-level changes. The model developed to explain that mass accumulation rates of carbonates diminished around 13.6 Ma (Mi3 Event) instead of 11.5 Ma (N12-N14 event), suggests that oceanic currents were controlling slope carbonate deposition on the Marion Plateau prior to the mid-Miocene, and that the mid Miocene event considerably increase their strength, hence reducing the amount of carbonate being deposited on slope sites. Moreover, by combining results from deep-sea oxygen isotopes with sea-level estimates based on coastal onlaps made during Leg 194, we constrain the amplitude of the N12-N14 sea-level fall to 90 meters. When integrating isotopic results from this study, this amplitude is lowered to 70 meters. A general conclusion of this work is that the mid Miocene climatic shift did impact carbonate systems, at least at the two locations studied. However, the nature of this response was highly dependant on the regional settings, in particular the presence of land mass (Malta) and the absence of a barrier to shelter from the effects of open ocean (Marion Plateau).
The present work is dealing with the first synthesis and characterisation of amphiphilic diblock copolymers bearing b-dicarbonyl (acetoacetoxy) chelating residues. Polymers were obtained by Group Transfer Polymerisation (GTP)/acetoacetylation and controlled radical polymerisation techniques (RAFT).Different micellar morphologies of poly(n-butyl methacrylate)-block-poly[2-(acetoacetoxy)ethyl methacrylate] (pBuMA-b-pAEMA) were observed in cyclohexane as a selective solvent. Depending on the block length ratio, either spherical, elliptical, or cylindrical micelles were formed. The density of the polymer chains at the core/corona interface is considerably higher as compared to any other strongly segregating system reported in the literature. It is demonstrated that there are H-bond interactions existing between acetoacetoxy groups, which increase the incompatibility between block segments. In addition, such interactions lead to the formation of secondary structures (such as b-sheets or globular structures) and larger superstructures in the micrometer length scale.Block copolymers were also used to solubilise metal ion salts of different geometries and oxidation states in organic media, in which are otherwise insoluble. Sterically stabilised colloidal hybrid materials are formed, i.e. monodisperse micelles having the metal ion salt incorporated in their core upon complexation with the ligating pAEMA block, whereas pBuMA forms the solvating corona responsible for stabilisation in solution. Systematic studies show that the aggregation behaviour is dependent on different factors, such as the tautomeric form of the beta-dicarbonyl ligand (keto/enol) as well as the nature and amount of added metal ion salt.
Movements of processive cytoskeletal motors are characterized by an interplay between directed motion along filament and diffusion in the surrounding solution. In the present work, these peculiar movements are studied by modeling them as random walks on a lattice. An additional subject of our studies is the effect of motor-motor interactions on these movements. In detail, four transport phenomena are studied: (i) Random walks of single motors in compartments of various geometries, (ii) stationary concentration profiles which build up as a result of these movements in closed compartments, (iii) boundary-induced phase transitions in open tube-like compartments coupled to reservoirs of motors, and (iv) the influence of cooperative effects in motor-filament binding on the movements. All these phenomena are experimentally accessible and possible experimental realizations are discussed.
The present work investigates the structure formation and wetting in two dimensional (2D) Langmuir monolayer phases in local thermodynamic equilibrium. A Langmuir monolayer is an isolated 2D system of surfactants at the air/water interface. It exhibits crystalline, liquid crystalline, liquid and gaseous phases differing in positional and/or orientational order. Permanent electric dipole moments of the surfactants lead to a long range repulsive interaction and to the formation of mesoscopic patterns. An interaction model is used describing the structure formation as a competition between short range attraction (bare line tension) and long range repulsion (surface potentials) on a scale Delta. Delta has the meaning of a dividing length between the short and long range interaction. In the present work the thermodynamic equilibrium conditions for the shape of two phase boundary lines (Young-Laplace equation) and three phase intersection points (Young′s condition) are derived and applied to describe experimental data: The line tension is measured by pendant droplet tensiometry. The bubble shape and size of 2D foams is calculated numerically and compared to experimental foams. Contact angles are measured by fitting numerical solutions of the Young-Laplace equation on micron scale. The scaling behaviour of the contact angle allows to measure a lower limit for Delta. Further it is discussed, whether in biological membranes wetting transitions are a way in order to control reaction kinetics. Studies performed in our group are discussed with respect to this question in the framework of the above mentioned theory. Finally the apparent violation of Gibbs′ phase rule in Langmuir monolayers (non-horizontal plateau of the surface pressure/area-isotherm, extended three phase coexistence region in one component systems) is investigated quantitatively. It has been found that the most probable explanation are impurities within the system whereas finite size effects or the influence of the long range electrostatics can not explain the order of magnitude of the effect.
In recent years, there has been a dramatic increase in available compute capacities. However, these “Grid resources” are rarely accessible in a continuous stream, but rather appear scattered across various machine types, platforms and operating systems, which are coupled by networks of fluctuating bandwidth. It becomes increasingly difficult for scientists to exploit available resources for their applications. We believe that intelligent, self-governing applications should be able to select resources in a dynamic and heterogeneous environment: Migrating applications determine a resource when old capacities are used up. Spawning simulations launch algorithms on external machines to speed up the main execution. Applications are restarted as soon as a failure is detected. All these actions can be taken without human interaction. A distributed compute environment possesses an intrinsic unreliability. Any application that interacts with such an environment must be able to cope with its failing components: deteriorating networks, crashing machines, failing software. We construct a reliable service infrastructure by endowing a service environment with a peer-to-peer topology. This “Grid Peer Services” infrastructure accommodates high-level services like migration and spawning, as well as fundamental services for application launching, file transfer and resource selection. It utilizes existing Grid technology wherever possible to accomplish its tasks. An Application Information Server acts as a generic information registry to all participants in a service environment. The service environment that we developed, allows applications e.g. to send a relocation requests to a migration server. The server selects a new computer based on the transmitted resource requirements. It transfers the application's checkpoint and binary to the new host and resumes the simulation. Although the Grid's underlying resource substrate is not continuous, we achieve persistent computations on Grids by relocating the application. We show with our real-world examples that a traditional genome analysis program can be easily modified to perform self-determined migrations in this service environment.
Our every-day experience is connected with different acoustical noise or music. Usually noise plays the role of nuisance in any communication and destroys any order in a system. Similar optical effects are known: strong snowing or raining decreases quality of a vision. In contrast to these situations noisy stimuli can also play a positive constructive role, e.g. a driver can be more concentrated in a presence of quiet music. Transmission processes in neural systems are of especial interest from this point of view: excitation or information will be transmitted only in the case if a signal overcomes a threshold. Dr. Alexei Zaikin from the Potsdam University studies noise-induced phenomena in nonlinear systems from a theoretical point of view. Especially he is interested in the processes, in which noise influences the behaviour of a system twice: if the intensity of noise is over a threshold, it induces some regular structure that will be synchronized with the behaviour of neighbour elements. To obtain such a system with a threshold one needs one more noise source. Dr. Zaikin has analyzed further examples of such doubly stochastic effects and developed a concept of these new phenomena. These theoretical findings are important, because such processes can play a crucial role in neurophysics, technical communication devices and living sciences.
In order to face the rapidly increasing need for computational resources of various scientific and engineering applications one has to think of new ways to make more efficient use of the worlds current computational resources. In this respect, the growing speed of wide area networks made a new kind of distributed computing possible: Metacomputing or (distributed) Grid computing. This is a rather new and uncharted field in computational science. The rapidly increasing speed of networks even outperforms the average increase of processor speed: Processor speeds double on average each 18 month whereas network bandwidths double every 9 months. Due to this development of local and wide area networks Grid computing will certainly play a key role in the future of parallel computing. This type of distributed computing, however, distinguishes from the traditional parallel computing in many ways since it has to deal with many problems not occurring in classical parallel computing. Those problems are for example heterogeneity, authentication and slow networks to mention only a few. Some of those problems, e.g. the allocation of distributed resources along with the providing of information about these resources to the application have been already attacked by the Globus software. Unfortunately, as far as we know, hardly any application or middle-ware software takes advantage of this information, since most parallelizing algorithms for finite differencing codes are implicitly designed for single supercomputer or cluster execution. We show that although it is possible to apply classical parallelizing algorithms in a Grid environment, in most cases the observed efficiency of the executed code is very poor. In this work we are closing this gap. In our thesis, we will - show that an execution of classical parallel codes in Grid environments is possible but very slow - analyze this situation of bad performance, nail down bottlenecks in communication, remove unnecessary overhead and other reasons for low performance - develop new and advanced algorithms for parallelisation that are aware of a Grid environment in order to generelize the traditional parallelization schemes - implement and test these new methods, replace and compare with the classical ones - introduce dynamic strategies that automatically adapt the running code to the nature of the underlying Grid environment. The higher the performance one can achieve for a single application by manual tuning for a Grid environment, the lower the chance that those changes are widely applicable to other programs. In our analysis as well as in our implementation we tried to keep the balance between high performance and generality. None of our changes directly affect code on the application level which makes our algorithms applicable to a whole class of real world applications. The implementation of our work is done within the Cactus framework using the Globus toolkit, since we think that these are the most reliable and advanced programming frameworks for supporting computations in Grid environments. On the other hand, however, we tried to be as general as possible, i.e. all methods and algorithms discussed in this thesis are independent of Cactus or Globus.
The colloidal systems are present everywhere in many varieties such as emulsions (liquid droplets dispersed in liquid), aerosols (liquid dispersed in gas), foam (gas in liquid), etc. Among several new methods for the preparation of colloids, the so-called miniemulsion technique has been shown to be one of the most promising. Miniemulsions are defined as stable emulsions consisting of droplets with a size of 50-500 nm by shearing a system containing oil, water, a surfactant, and a highly water insoluble compound, the so-called hydrophobe 1. In the first part of this work, dynamic crystallization and melting experiments are described which were performed in small, stable and narrowly distributed nanodroplets (confined systems) of miniemulsions. Both regular and inverse systems were examined, characterizing, first, the crystallization of hexadecane, secondly, the crystallization of ice. It was shown for both cases that the temperature of crystallization in such droplets is significantly decreased (or the required undercooling is increased) as compared to the bulk material. This was attributed to a very effective suppression of heterogeneous nucleation. It was also found that the required undercooling depends on the nanodroplet size: with decreasing droplet size the undercooling increases. 2. It is shown that the temperature of crystallization of other n-alkanes in nanodroplets is also significantly decreased as compared to the bulk material due to a very effective suppression of heterogeneous nucleation. A very different behavior was detected between odd and even alkanes. In even alkanes, the confinement in small droplets changes the crystal structure from a triclinic (as seen in bulk) to an orthorhombic structure, which is attributed to finite size effects inside the droplets. An intermediate metastable rotator phase is of less relevance for the miniemulsion droplets than in the bulk. For odd alkanes, only a strong temperature shift compared to the bulk system is observed, but no structure change. A triclinic structure is formed both in bulk and in miniemulsion droplets. 3. In the next part of the thesis it is shown how miniemulsions could be successfully applied in the development of materials with potential application in pharmaceutical and medical fields. The production of cross-linked gelatin nanoparticles is feasible. Starting from an inverse miniemulsion, the softness of the particles can be controlled by varying the initial concentration, amount of cross-link agent, time of cross-linking, among other parameters. Such particles show a thermo-reversible effect, e.g. the particles swell in water above 37 °C and shrink below this temperature. Above 37 °C the chains loose the physical cross-linking, however the particles do not loose their integrity, because of the chemical cross-linking. Those particles have potential use as drug carriers, since gelatin is a natural polymer derived from collagen. 4. The cross-linked gelatin nanoparticles have been used for the biomineralization of hydroxyapatite (HAP), a biomineral, which is the major constituent of our bones. The biomineralization of HAP crystals within the gelatin nanoparticles results in a hybrid material, which has potential use as a bone repair material. 5. In the last part of this work we have shown that layers of conjugated semiconducting polymers can be deposited from aqueous dispersion prepared by the miniemulsion process. Dispersions of particles of different conjugated semiconducting polymers such as a ladder-type poly(para-phenylene) and several soluble derivatives of polyfluorene could be prepared with well-controlled particle sizes ranging between 70 - 250 nm. Layers of polymer blends were prepared with controlled lateral dimensions of phase separation on sub-micrometer scales, utilizing either a mixture of single component nanoparticles or nanoparticles containing two polymers. From the results of energy transfer it is demonstrated that blending two polymers in the same particle leads to a higher efficiency due to the better contact between the polymers. Such an effect is of great interest for the fabrication of opto-electronic devices such as light emitting diodes with nanometer size emitting points and solar cells comprising of blends of electron donating and electron accepting polymers.
Studies of the role of disturbance in vegetation or ecosystems showed that disturbances are an essential and intrinsic element of ecosystems that contribute substantially to ecosystem health, to structural diversity of ecosystems and to nutrient cycling at the local as well as global level. Fire as a grassland, bush or forest fire is a special disturbance agent, since it is caused by biotic as well abiotic environmental factors. Fire affects biogeochemical cycles and plays an important role in atmospheric chemistry by releasing climate-sensitive trace gases and aerosols, and thus in the global carbon cycle by releasing approximately 3.9 Gt C p.a. through biomass burning. A combined model to describe effects and feedbacks between fire and vegetation became relevant as changes in fire regimes due to land use and land management were observed and the global dimension of biomass burnt as an important carbon flux to the atmosphere, its influence on atmospheric chemistry and climate as well as vegetation dynamics were emphasized. The existing modelling approaches would not allow these investigations. As a consequence, an optimal set of variables that best describes fire occurrence, fire spread and its effects in ecosystems had to be defined, which can simulate observed fire regimes and help to analyse interactions between fire and vegetation dynamics as well as to allude to the reasons behind changing fire regimes. Especially, dynamic links between vegetation, climate and fire processes are required to analyse dynamic feedbacks and effects of changes of single environmental factors. This led us to the point, where new fire models had to be developed that would allow the investigations, mentioned above, and could help to improve our understanding of the role of fire in global ecology. In conclusion of the thesis, one can state that moisture conditions, its persistence over time and fuel load are the important components that describe global fire pattern. If time series of a particular region are to be reproduced, specific ignition sources, fire-critical climate conditions and vegetation composition become additional determinants. Vegetation composition changes the level of fire occurrence and spread, but has limited impact on the inter-annual variability of fire. The importance to consider the full range of major fire processes and links to vegetation dynamics become apparent under climate change conditions. Increases in climate-dependent length of fire season does not automatically imply increases in biomass burnt, it can be buffered or accelerated by changes in vegetation productivity. Changes in vegetation composition as well as enhanced vegetation productivity can intensify changes in fire and lead to even more fire-related emissions. --- Anmerkung: Die Autorin ist Trägerin des von der Mathematisch-Naturwissenschaftlichen Fakultät der Universität Potsdam vergebenen Michelson-Preises für die beste Promotion des Jahres 2002/2003.
In this thesis, I investigated the factors influencing the growth and vertical distribution of planktonic algae in extremely acidic mining lakes (pH 2-3). In the focal study site, Lake 111 (pH 2.7; Lusatia, Germany), the chrysophyte, Ochromonas sp., dominates in the upper water strata and the chlorophyte, Chlamydomonas sp., in the deeper strata, forming a pronounced deep chlorophyll maximum (DCM). Inorganic carbon (IC) limitation influenced the phototrophic growth of Chlamydomonas sp. in the upper water strata. Conversely, in deeper strata, light limited its phototrophic growth. When compared with published data for algae from neutral lakes, Chlamydomonas sp. from Lake 111 exhibited a lower maximum growth rate, an enhanced compensation point and higher dark respiration rates, suggesting higher metabolic costs due to the extreme physico-chemical conditions. The photosynthetic performance of Chlamydomonas sp. decreased in high-light-adapted cells when IC limited. In addition, the minimal phosphorus (P) cell quota was suggestive of a higher P requirement under IC limitation. Subsequently, it was shown that Chlamydomonas sp. was a mixotroph, able to enhance its growth rate by taking up dissolved organic carbon (DOC) via osmotrophy. Therefore, it could survive in deeper water strata where DOC concentrations were higher and light limited. However, neither IC limitation, P availability nor in situ DOC concentrations (bottom-up control) could fully explain the vertical distribution of Chlamydomonas sp. in Lake 111. Conversely, when a novel approach was adopted, the grazing influence of the phagotrophic phototroph, Ochromonas sp., was found to exert top-down control on its prey (Chlamydomonas sp.) reducing prey abundance in the upper water strata. This, coupled with the fact that Chlamydomonas sp. uses DOC for growth, leads to a pronounced accumulation of Chlamydomonas sp. cells at depth; an apparent DCM. Therefore, grazing appears to be the main factor influencing the vertical distribution of algae observed in Lake 111. The knowledge gained from this thesis provides information essential for predicting the effect of strategies to neutralize the acidic mining lakes on the food-web.
One of the rules-of-thumb of colloid and surface physics is that most surfaces are charged when in contact with a solvent, usually water. This is the case, for instance, in charge-stabilized colloidal suspensions, where the surface of the colloidal particles are charged (usually with a charge of hundreds to thousands of e, the elementary charge), monolayers of ionic surfactants sitting at an air-water interface (where the water-loving head groups become charged by releasing counterions), or bilayers containing charged phospholipids (as cell membranes). In this work, we look at some model-systems that, although being a simplified version of reality, are expected to capture some of the physical properties of real charged systems (colloids and electrolytes). We initially study the simple double layer, composed by a charged wall in the presence of its counterions. The charges at the wall are smeared out and the dielectric constant is the same everywhere. The Poisson-Boltzmann (PB) approach gives asymptotically exact counterion density profiles around charged objects in the weak-coupling limit of systems with low-valent counterions, surfaces with low charge density and high temperature (or small Bjerrum length). Using Monte Carlo simulations, we obtain the profiles around the charged wall and compare it with both Poisson-Boltzmann (in the low coupling limit) and the novel strong coupling (SC) theory in the opposite limit of high couplings. In the latter limit, the simulations show that the SC leads in fact to asymptotically correct density profiles. We also compare the Monte Carlo data with previously calculated corrections to the Poisson-Boltzmann theory. We also discuss in detail the methods used to perform the computer simulations. After studying the simple double layer in detail, we introduce a dielectric jump at the charged wall and investigate its effect on the counterion density distribution. As we will show, the Poisson-Boltzmann description of the double layer remains a good approximation at low coupling values, while the strong coupling theory is shown to lead to the correct density profiles close to the wall (and at all couplings). For very large couplings, only systems where the difference between the dielectric constants of the wall and of the solvent is small are shown to be well described by SC. Another experimentally relevant modification to the simple double layer is to make the charges at the plane discrete. The counterions are still assumed to be point-like, but we constraint the distance of approach between ions in the plane and counterions to a minimum distance D. The ratio between D and the distance between neighboring ions in the plane is, as we will see, one of the important quantities in determining the influence of the discrete nature of the charges at the wall over the density profiles. Another parameter that plays an important role, as in the previous case, is the coupling as we will demonstrate, systems with higher coupling are more subject to discretization effects than systems with low coupling parameter. After studying the isolated double layer, we look at the interaction between two double layers. The system is composed by two equally charged walls at distance d, with the counterions confined between them. The charge at the walls is smeared out and the dielectric constant is the same everywhere. Using Monte-Carlo simulations we obtain the inter-plate pressure in the global parameter space, and the pressure is shown to be negative (attraction) at certain conditions. The simulations also show that the equilibrium plate separation (where the pressure changes from attractive to repulsive) exhibits a novel unbinding transition. We compare the Monte Carlo results with the strong-coupling theory, which is shown to describe well the bound states of systems with moderate and high couplings. The regime where the two walls are very close to each other is also shown to be well described by the SC theory. Finally, Using a field-theoretic approach, we derive the exact low-density ("virial") expansion of a binary mixture of positively and negatively charged hard spheres (two-component hard-core plasma, TCPHC). The free energy obtained is valid for systems where the diameters d_+ and d_- and the charge valences q_+ and q_- of positive and negative ions are unconstrained, i.e., the same expression can be used to treat dilute salt solutions (where typically d_+ ~ d_- and q_+ ~ q_-) as well as colloidal suspensions (where the difference in size and valence between macroions and counterions can be very large). We also discuss some applications of our results.
Today, analytical chemistry does not longer consist of only the big measuring devices and methods which are time consuming and expensive, which can furthermore only be handled by the qualified staff and in addition the results can also only be evaluated by this qualified staff. Usually, this technique, which shall be described in the following as 'classic analytic measuring technique', requires also rooms equipped especially and often a relative big quantity of the test compounds which should be prepared especially. Beside this classic analytic measuring technique, limited on definite substance groups and requests, a new measuring technique has gained acceptance particularly within the last years, which one can often be used by a layman, too. Often the new measuring technique has very little pieces of equipment. The needed sample volumes are also small and a special sample preparation isn't required. In addition, the new measuring instruments are simple to handle. They are cheap both in their production and in the use and they permit even a continuous measurement recording usually. Numerous of this new measuring instruments base on the research in the field of Biosensorik during the last 40 years. Since Clark and Lyon in the year 1962 were able to measure glucose with a simple oxygen electrode, completed by an enzyme the development of the new measuring technique did not have to be held back any longer. Biosensors, special pickups which consists of a combination from a biological component (permits a specific recognition of the analyte also without purification of the sample previously) and a physical pickup (convert the primary physicochemical effect into an electronically measurable signal), conquered the market. In the context of this thesis different tyrosinasesensors were developed which fulfilling the various requests, depending on origin and features of the used tyrosinase. One of the tyrosinasesensors for example was used for quantification of phenolic compounds in river and sea water and the results could correlated very well with the corresponding DIN-test for the determination of phenolic compounds. An other developed tyrosinasesensor showed a very high sensitiveness for catecholamines, substances which are of special importance in the medical diagnostics. In addition, the investigations of two different tyrosinases, which were carried out also in the context of this thesis, have shown, that a special tyrosinase (tyrosinase from Streptomyces antibioticus) will be the better choice as tyrosinase from Agaricus bisporus, which is used in the area of biosensor research till now, if one wants to develop in future even more sensitive tyrosinasesensors. Furthermore, first successes became reached on a molecular biological field, the production of tyrosinasemutants with special, before well-considered features. These successes can be used to develop a new generation of tyrosinasesensors, tyrosinasesensors in which tyrosinase can be bound directionally both to the corresponding physical pickup or also to another enzyme. From this one expects to achieve ways minimized which the substance to be determined (or whose product) otherwise must cover. Finally, this should result in an clearly visible increase of sensitivity of the Biosensor.
Combined structural and magnetotelluric investigation across the West Fault Zone in northern Chile
(2002)
The characterisation of the internal architecture of large-scale fault zones is usually restricted to the outcrop-based investigation of fault-related structural damage on the Earth's surface. A method to obtain information on the downward continuation of a fault is to image the subsurface electrical conductivity structure. This work deals with such a combined investigation of a segment of the West Fault, which itself is a part of the more than 2000 km long trench-linked Precordilleran Fault System in the northern Chilean Andes. Activity on the fault system lasted from Eocene to Quaternary times. In the working area (22°04'S, 68°53'W), the West Fault exhibits a clearly defined surface trace with a constant strike over many tens of kilometers. Outcrop condition and morphology of the study area allow ideally for a combination of structural geology investigation and magnetotelluric (MT) / geomagnetic depth sounding (GDS) experiments. The aim was to achieve an understanding of the correlation of the two methods and to obtain a comprehensive view of the West Fault's internal architecture. Fault-related brittle damage elements (minor faults and slip-surfaces with or without striation) record prevalent strike-slip deformation on subvertically oriented shear planes. Dextral and sinistral slip events occurred within the fault zone and indicate reactivation of the fault system. Youngest deformation increments mapped in the working area are extensional and the findings suggest a different orientation of the extension axes on either side of the fault. Damage element density increases with approach to the fault trace and marks an approximately 1000 m wide damage zone around the fault. A region of profound alteration and comminution of rocks, about 400 m wide, is centered in the damage zone. Damage elements in this central part are predominantly dipping steeply towards the east (70-80°). Within the same study area, the electrical conductivity image of the subsurface was measured along a 4 km long MT/GDS profile. This main profile trends perpendicular to the West Fault trace. The MT stations of the central 2 km were 100 m apart from each other. A second profile with 300 m site spacing and 9 recording sites crosses the fault a few kilometers away from the main study area. Data were recorded in the frequency range from 1000 Hz to 0.001 Hz with four real time instruments S.P.A.M. MkIII. The GDS data reveal the fault zone for both profiles at frequencies above 1 Hz. Induction arrows indicate a zone of enhanced conductivity several hundred meters wide, that aligns along the WF strike and lies mainly on the eastern side of the surface trace. A dimensionality analysis of the MT data justifies a two dimensional model approximation of the data for the frequency range from 1000 Hz to 0.1 Hz. For this frequency range a regional geoelectric strike parallel to the West Fault trace could be recovered. The data subset allows for a resolution of the conductivity structure of the uppermost crust down to at least 5 km. Modelling of the MT data is based on an inversion algorithm developed by Mackie et al. (1997). The features of the resulting resistivity models are tested for their robustness using empirical sensitivity studies. This involves variation of the properties (geometry, conductivity) of the anomalies, the subsequent calculation of forward or constrained inversion models and check for consistency of the obtained model results with the data. A fault zone conductor is resolved on both MT profiles. The zones of enhanced conductivity are located to the east of the West Fault surface trace. On the dense MT profile, the conductive zone is confined to a width of about 300 m and the anomaly exhibits a steep dip towards the east (about 70°). Modelling implies that the conductivity increase reaches to a depth of at least 1100 m and indicates a depth extent of less than 2000 m. Further conductive features are imaged but their geometry is less well constrained. The fault zone conductors of both MT profiles coincide in position with the alteration zone. For the dense profile, the dip of the conductive anomaly and the dip of the damage elements of the central part of the fault zone correlate. This suggests that the electrical conductivity enhancement is causally related to a mesh of minor faults and fractures, which is a likely pathway for fluids. The interconnected rock-porosity that is necessary to explain the observed conductivity enhancement by means of fluids is estimated on the basis of the salinity of several ground water samples (Archie's Law). The deeper the source of the water sample, the more saline it is due to longer exposure to fluid-rock interaction and the lower is the fluid's resistivity. A rock porosity in the range of 0.8% - 4% would be required at a depth of 200 m. That indicates that fluids penetrating the damaged fault zone from close to the surface are sufficient to explain the conductivity anomalies. This is as well supported by the preserved geochemical signature of rock samples in the alteration zone. Late stage alteration processes were active in a low temperature regime (<95°C) and the involvement of ascending brines from greater depth is not indicated. The limited depth extent of the fault zone conductors is a likely result of sealing and cementation of the fault fracture mesh due to dissolution and precipitation of minerals at greater depth and increased temperature. Comparison of the results of the apparently inactive West Fault with published studies on the electrical conductivity structure of the currently active San Andreas Fault, suggests that the depth extent and conductivity of the fault zone conductor may be correlated to fault activity. Ongoing deformation will keep the fault/fracture mesh permeable for fluids and impede cementation and sealing of fluid pathways.
Semi-arid areas are, due to their climatic setting, characterized by small water resources. An increasing water demand as a consequence of population growth and economic development as well as a decreasing water availability in the course of possible climate change may aggravate water scarcity in future, which often exists already for present-day conditions in these areas. Understanding the mechanisms and feedbacks of complex natural and human systems, together with the quantitative assessment of future changes in volume, timing and quality of water resources are a prerequisite for the development of sustainable measures of water management to enhance the adaptive capacity of these regions. For this task, dynamic integrated models, containing a hydrological model as one component, are indispensable tools. The main objective of this study is to develop a hydrological model for the quantification of water availability in view of environmental change over a large geographic domain of semi-arid environments. The study area is the Federal State of Ceará (150 000 km2) in the semi-arid north-east of Brazil. Mean annual precipitation in this area is 850 mm, falling in a rainy season with duration of about five months. Being mainly characterized by crystalline bedrock and shallow soils, surface water provides the largest part of the water supply. The area has recurrently been affected by droughts which caused serious economic losses and social impacts like migration from the rural regions. The hydrological model Wasa (Model of Water Availability in Semi-Arid Environments) developed in this study is a deterministic, spatially distributed model being composed of conceptual, process-based approaches. Water availability (river discharge, storage volumes in reservoirs, soil moisture) is determined with daily resolution. Sub-basins, grid cells or administrative units (municipalities) can be chosen as spatial target units. The administrative units enable the coupling of Wasa in the framework of an integrated model which contains modules that do not work on the basis of natural spatial units. The target units mentioned above are disaggregated in Wasa into smaller modelling units within a new multi-scale, hierarchical approach. The landscape units defined in this scheme capture in particular the effect of structured variability of terrain, soil and vegetation characteristics along toposequences on soil moisture and runoff generation. Lateral hydrological processes at the hillslope scale, as reinfiltration of surface runoff, being of particular importance in semi-arid environments, can thus be represented also within the large-scale model in a simplified form. Depending on the resolution of available data, small-scale variability is not represented explicitly with geographic reference in Wasa, but by the distribution of sub-scale units and by statistical transition frequencies for lateral fluxes between these units. Further model components of Wasa which respect specific features of semi-arid hydrology are: (1) A two-layer model for evapotranspiration comprises energy transfer at the soil surface (including soil evaporation), which is of importance in view of the mainly sparse vegetation cover. Additionally, vegetation parameters are differentiated in space and time in dependence on the occurrence of the rainy season. (2) The infiltration module represents in particular infiltration-excess surface runoff as the dominant runoff component. (3) For the aggregate description of the water balance of reservoirs that cannot be represented explicitly in the model, a storage approach respecting different reservoirs size classes and their interaction via the river network is applied. (4) A model for the quantification of water withdrawal by water use in different sectors is coupled to Wasa. (5) A cascade model for the temporal disaggregation of precipitation time series, adapted to the specific characteristics of tropical convective rainfall, is applied for the generating rainfall time series of higher temporal resolution. All model parameters of Wasa can be derived from physiographic information of the study area. Thus, model calibration is primarily not required. Model applications of Wasa for historical time series generally results in a good model performance when comparing the simulation results of river discharge and reservoir storage volumes with observed data for river basins of various sizes. The mean water balance as well as the high interannual and intra-annual variability is reasonably represented by the model. Limitations of the modelling concept are most markedly seen for sub-basins with a runoff component from deep groundwater bodies of which the dynamics cannot be satisfactorily represented without calibration. Further results of model applications are: (1) Lateral processes of redistribution of runoff and soil moisture at the hillslope scale, in particular reinfiltration of surface runoff, lead to markedly smaller discharge volumes at the basin scale than the simple sum of runoff of the individual sub-areas. Thus, these processes are to be captured also in large-scale models. The different relevance of these processes for different conditions is demonstrated by a larger percentage decrease of discharge volumes in dry as compared to wet years. (2) Precipitation characteristics have a major impact on the hydrological response of semi-arid environments. In particular, underestimated rainfall intensities in the rainfall input due to the rough temporal resolution of the model and due to interpolation effects and, consequently, underestimated runoff volumes have to be compensated in the model. A scaling factor in the infiltration module or the use of disaggregated hourly rainfall data show good results in this respect. The simulation results of Wasa are characterized by large uncertainties. These are, on the one hand, due to uncertainties of the model structure to adequately represent the relevant hydrological processes. On the other hand, they are due to uncertainties of input data and parameters particularly in view of the low data availability. Of major importance is: (1) The uncertainty of rainfall data with regard to their spatial and temporal pattern has, due to the strong non-linear hydrological response, a large impact on the simulation results. (2) The uncertainty of soil parameters is in general of larger importance on model uncertainty than uncertainty of vegetation or topographic parameters. (3) The effect of uncertainty of individual model components or parameters is usually different for years with rainfall volumes being above or below the average, because individual hydrological processes are of different relevance in both cases. Thus, the uncertainty of individual model components or parameters is of different importance for the uncertainty of scenario simulations with increasing or decreasing precipitation trends. (4) The most important factor of uncertainty for scenarios of water availability in the study area is the uncertainty in the results of global climate models on which the regional climate scenarios are based. Both a marked increase or a decrease in precipitation can be assumed for the given data. Results of model simulations for climate scenarios until the year 2050 show that a possible future change in precipitation volumes causes a larger percentage change in runoff volumes by a factor of two to three. In the case of a decreasing precipitation trend, the efficiency of new reservoirs for securing water availability tends to decrease in the study area because of the interaction of the large number of reservoirs in retaining the overall decreasing runoff volumes.
Deep convection is an essential part of the circulation in the North Atlantic Ocean. It influences the northward heat transport achieved by the thermohaline circulation. Understanding its stability and variability is therefore necessary for assessing climatic changes in the area of the North Atlantic. This thesis aims at improving the conceptual understanding of the stability and variability of deep convection. Observational data from the Labrador Sea show phases with and without deep convection. A simple two-box model is fitted to these data. The results suggest that the Labrador Sea has two coexisting stable states, one with regular deep convection and one without deep convection. This bistability arises from a positive salinity feedback that is due to the net freshwater input into the surface layer. The convecting state can easily become unstable if the mean forcing shifts to warmer or less saline conditions. The weather-induced variability of the external forcing is included into the box model by adding a stochastic forcing term. It turns out that deep convection is then switched "on" and "off" frequently. The mean residence time in either state is a measure of its stochastic stability. The stochastic stability depends smoothly on the forcing parameters, in contrast to the deterministic (non-stochastic) stability which may change abruptly. The mean and the variance of the stochastic forcing both have an impact on the frequency of deep convection. For instance, a decline in convection frequency due to a surface freshening may be compensated for by an increased heat flux variability. With a further simplified box model some stochastic stability features are studied analytically. A new effect is described, called wandering monostability: even if deep convection is not a stable state due to changed forcing parameters, the stochastic forcing can still trigger convection events frequently. The analytical expressions explicitly show how wandering monostability and other effects depend on the model parameters. This dependence is always exponential for the mean residence times, but for the probability of long nonconvecting phases it is exponential only if this probability is small. It is to be expected that wandering monostability is relevant in other parts of the climate system as well. All in all, the results demonstrate that the stability of deep convection in the Labrador Sea reacts very sensitively to the forcing. The presence of variability is crucial for understanding this sensitivity. Small changes in the forcing can already significantly lower the frequency of deep convection events, which presumably strongly affects the regional climate. ----Anmerkung: Der Autor ist Träger des durch die Physikalische Gesellschaft zu Berlin vergebenen Carl-Ramsauer-Preises 2003 für die jeweils beste Dissertation der vier Universitäten Freie Universität Berlin, Humboldt-Universität zu Berlin, Technische Universität Berlin und Universität Potsdam.
Motivated by recent proposals on the experimental detectability of quantum gravity effects, the present thesis investigates assumptions and methods which might be used for the prediction of such effects within the framework of loop quantum gravity. To this end, a scalar field coupled to gravity is considered as a model system. Starting from certain assumptions about the dynamics of the coupled gravity-matter system, a quantum theory for the scalar field is proposed. Then, assuming that the gravitational field is in a semiclassical state, a "QFT on curved space-time limit" of this theory is defined. In contrast to ordinary quantum field theory on curved space-time however, in this limit the theory describes a quantum scalar field propagating on a (classical) random lattice. Then, methods to obtain the low energy limit of such a lattice theory, especially regarding the resulting modified dispersion relations, are discussed and applied to simple model systems. Finally, under certain simplifying assumptions, using the methods developed before as well as a specific class of semiclassical states, corrections to the dispersion relations for the scalar and the electromagnetic field are computed within the framework of loop quantum gravity. These calculations are of preliminary character, as many assumptions enter whose validity remains to be studied more thoroughly. However they exemplify the problems and possibilities of making predictions based on loop quantum gravity that are in principle testable by experiment.
Structural and spectroscopical study of crystals of 1,3,4-oxadiazole derivatives at high pressure
(2002)
In recent years the search for new materials of technological interest has given new impulses to the study of organic compounds. Organic substances possess a great number of advantages such as the possibility to adjust their properties for a given purpose by different chemical and physical techniques in the preparation process. Oxadiazole derivatives are interesting due to their use as material for light emitting diodes (LED) as well as scintillators. The physical properties of a solid depend on its structure. Different structures induce different intra- and intermolecular interactions. An advantageous method to modify the intra- as well as the intermolecular interactions of a given substance is the application of high pressure. Furthermore, using this method the chemical features of the compound are not influenced. We have investigated the influence of high pressure and high temperature on the super-molecular structure of several oxadiazole derivatives in crystalline state. From the results of this investigation an equation of state for these crystals was determined. Furthermore, the spectroscopical features of these materials under high pressure were characterized.