Refine
Has Fulltext
- yes (156) (remove)
Year of publication
- 2010 (156) (remove)
Document Type
- Doctoral Thesis (47)
- Conference Proceeding (24)
- Monograph/Edited Volume (23)
- Postprint (22)
- Article (21)
- Preprint (7)
- Review (6)
- Other (2)
- Working Paper (2)
- Habilitation Thesis (1)
Language
- English (156) (remove)
Keywords
- middleware (4)
- Arrayseismologie (2)
- Aspektorientierte Softwareentwicklung (2)
- Betriebssysteme (2)
- Constraint Solving (2)
- Cosmogenic nuclides (2)
- Deduction (2)
- Erdbeben (2)
- Erdbebenkatalog (2)
- Erdbebenschwarm 2008/09 (2)
- Gelatine (2)
- Germany (2)
- InSAR (2)
- Integration (2)
- Klimawandel (2)
- Kosmogene Nuklide (2)
- Logic Programming (2)
- Logics (2)
- Middleware (2)
- Optimization (2)
- Planing (2)
- Standorteffekte (2)
- Systemsoftware (2)
- Virtuelle Maschinen (2)
- Vogtland/West Bohemia (2)
- Vogtland/Westböhmen (2)
- array seismology (2)
- earthquake catalog (2)
- earthquake swarm 2008/09 (2)
- eye movements (2)
- freshwater (2)
- operating systems (2)
- reading (2)
- seismic noise (2)
- site effects (2)
- systems software (2)
- tax reform (2)
- total internal reflection fluorescence microscopy (2)
- virtual machines (2)
- (Alters-) Datierungen (1)
- 3D city models (1)
- 3D geovisualisation (1)
- 3D-Geovisualisierung (1)
- 3D-Stadtmodelle (1)
- A-bar-movement (1)
- Abbau (1)
- Accentuation (1)
- Ackerschmalwand (1)
- Agrammatismus (1)
- Akan (1)
- Aktienmarkt (1)
- Alborz (1)
- Antipersistence (1)
- Arabidopsis (1)
- Arbeitsethik (1)
- Armed Conflicts (1)
- Artenreichtum (1)
- Aspekt (1)
- Asynchrone Schaltung (1)
- Asynchronous circuit (1)
- Atomic Force Microscope (1)
- Attributanalyse (1)
- Aufgabenerfüllung (1)
- Außenwirtschaftstheorie (1)
- Bayesian Value-at-Risk (1)
- Bayesianischer Value-at-Risk (1)
- Bildbearbeitung (1)
- Biogenic amine (1)
- Biomaterial (1)
- Biomaterialien (1)
- Biosensor (1)
- Blaulichtsensoren (1)
- Blickbewegungsmessung (1)
- Blue-light sensors (1)
- Bodenwasser (1)
- Bottom-up (1)
- Bulgarian (1)
- Bulgarisch (1)
- Business Process Models (1)
- Bürokratisierung (1)
- CSC (1)
- Cascading (1)
- Central Banking Policy (1)
- Climate change (1)
- Coccinelle (1)
- Collective Identity (1)
- Colombia (1)
- Compacton (1)
- Compliance (1)
- Composite (1)
- Consolidation (1)
- Consumption (1)
- Controller-Resynthese (1)
- Crime mapping (1)
- Cross Contamination (1)
- Cultural Integration (1)
- Curriculum Framework (1)
- Cyclic linearization (1)
- Cytochrom c (1)
- Czech (1)
- Datenanalyse (1)
- Datenintegration (1)
- Decentralization (1)
- Deutschland (1)
- Dezentralisierung (1)
- Dienstkomposition (1)
- Dienstplattform (1)
- Distribution functions with upper bound (1)
- Donovani (1)
- Druck-Temperatur Bedingungen (1)
- Dynamic Type System (1)
- Dynamische Typ Systeme (1)
- EEG (1)
- Earthquake (1)
- Earthquake magnitude (1)
- East African Community (1)
- East Antarctica (1)
- Einnahmenautonomie (1)
- Eisdynamik (1)
- Eisen (1)
- Eklogite (1)
- Elburs (1)
- Electroactive material (1)
- Elektroaktive Materialien (1)
- Elektrochemie (1)
- Elektrospinnen (1)
- Elliptizität (1)
- Energiehaushalt (1)
- Epidemiologie (1)
- Epidemiology (1)
- Erdbeben-Magnitude (1)
- Erzählung (1)
- Estimation for branching processes (1)
- European Union (1)
- European values education (1)
- Europäische Union (1)
- Europäische Werteerziehung (1)
- Evaluierung der Steuerpolitik (1)
- Exhumationsraten (1)
- Expenditure Assignment (1)
- Expert Elicitation (1)
- Expertenbefragung (1)
- Exploration (1)
- Faser (1)
- Fault interaction (1)
- Fehlersuche (1)
- Feller Diffusionsprozesse (1)
- Feller diffusion processes (1)
- Fernerkundung (1)
- Ferroelektret (1)
- Feuer (1)
- Finanzzuweisungen (1)
- Finite Differenzen (1)
- Fiscal Federalism (1)
- Fiskalischer Föderalismus (1)
- Flexion (1)
- Flood frequency analysis (1)
- Flood regionalisation (1)
- Fluid-Schmelze Wechselwirkung (1)
- Focus (1)
- Foraging (1)
- Forschungskolleg (1)
- Frequenz (1)
- G-protein-coupled receptor (1)
- GIS (1)
- GITEWS (1)
- GOAT (1)
- GPS (1)
- Gelatin (1)
- Gene Ontology (1)
- Genotypisierung (1)
- Geomagnetism (1)
- Georadar (1)
- German (1)
- Geschäftsprozessmodelle (1)
- Gewässer (1)
- Ghrelin (1)
- Gitterdynamik (1)
- Glaciers (1)
- Gletscher (1)
- Globaler Wasserkreislauf (1)
- Granular chain (1)
- Graphen (1)
- Growth (1)
- HDAC (1)
- HP-HT Experimente (1)
- HP-HT experiments (1)
- Hamilton (1)
- Hamiltonian (1)
- Hasso Plattner Institute (1)
- Hasso-Plattner-Institut (1)
- Herodotos (1)
- High-Level Synthesis (1)
- Himalaja (Kaghan) (1)
- Himalaya (1)
- Himalaya (Kaghan Valley) (1)
- History of pattern occurrences (1)
- Hochwasserregionalisierung (1)
- Hochwasserstatistik (1)
- Human Rights (1)
- Hungerhormon (1)
- Hydrogele (1)
- Hydrogeophysics (1)
- Hydrogeopyhsik (1)
- Hydrolgische Modellierung (1)
- Hydroxyapatit (1)
- Hydroxyapatite (1)
- Immigration (1)
- Immobilization (1)
- In-Situ-Analyse (1)
- Information Extraction (1)
- Information structure (1)
- Informationsextraktion (1)
- Intergovernmental transfers (1)
- Intervention effects (1)
- Investment (1)
- Iran (1)
- Judaism (1)
- Judentum (1)
- Klausurtagung (1)
- Kobalt (1)
- Kollektive Identität (1)
- Kolumbien (1)
- Komplementsatzverstehen (1)
- Komposite (1)
- Korrelation (1)
- LHCII (1)
- Landesgröße (1)
- Lattice dynamics (1)
- Laufzeitanalyse (1)
- Lehrevaluation (1)
- Lesen (1)
- Lexikalischer Zugriff (1)
- Linked Data (1)
- Länder (1)
- MCMC Simulation (1)
- MHD (1)
- Magnetfeld-Satellit (1)
- Magnetic field measurements (1)
- Magnetische Feldmessungen (1)
- Magnetometer-Kalibrierung (1)
- Mehrtyp-Verzweigungsprozesse (1)
- Menschenrechte (1)
- Metadatenentdeckung (1)
- Metadatenqualität (1)
- Microcontact Printing (1)
- Microtus (1)
- Mikrosimulation (1)
- Ministerialverwaltung (1)
- Model Weighting (1)
- Modelgewichtung (1)
- Modellierung (1)
- Molecular modeling (1)
- Molekulare Modellierung (1)
- Molekülmodelle (1)
- Monte Carlo testing (1)
- Multiprocessor (1)
- Multiprozessor (1)
- Multitype branching processes (1)
- Muster (1)
- Musterabgleich (1)
- NLO (1)
- NW Iran (1)
- Nahrungsfette (1)
- Nahrungssuche (1)
- Nanofluidik (1)
- Nanopartikel (1)
- Negation (1)
- Neo-Tethys (1)
- Neue ökonomische Geographie (1)
- Next Generation Network (1)
- Ngizim (1)
- Oberflächenkräfte (1)
- Object-Oriented Programming (1)
- Objekt-Orientiertes Programmieren (1)
- Optimality (1)
- Optimalität (1)
- Ordinary Differential Equation model (1)
- Ostantarktis (1)
- P(VDF-TrFE) (1)
- PDLC (1)
- PDMS (1)
- Parallel Programming (1)
- Parallelrechner (1)
- Parametrices (1)
- Pattern Matching (1)
- Patterns (1)
- Persistenz (1)
- Petrinetz (1)
- Ph.D. Retreat (1)
- Phenotyping (1)
- Phosphorylation Site (1)
- Photoelektronenmikroskopie (1)
- Photosynthese (1)
- Photosynthesis (1)
- Physical Network (1)
- Physikalische Quervernetzung (1)
- Phytoplankton (1)
- Phänologie (1)
- Phänotypisierung (1)
- Piezoelektrizität (1)
- Poland (1)
- Polen (1)
- Policy Enforcement (1)
- Politisierung (1)
- Polymer-Netzwerke (1)
- Polymerfilm (1)
- Polymerschaum (1)
- Polynia (1)
- Predation (1)
- Prediction Markets (1)
- Probabilistic regional envelope curves (1)
- Probabilistische Regionale Hüllkurven (1)
- Programmierung (1)
- Prolog (1)
- Property Prices (1)
- Prosody-syntax interface (1)
- Protein-Engineering (1)
- Proteom (1)
- Prädation (1)
- Psalm 58 (1)
- Pseudo-Differentialoperatoren (1)
- Pseudo-differential operators (1)
- Psycholinguistik (1)
- QPM (1)
- Quellenparameter (1)
- RNA (1)
- Rashba effect (1)
- Rashba-Effekt (1)
- Reconfigurable (1)
- Remote sensing (1)
- Research School (1)
- Revenue Autonomy (1)
- Rheologie (1)
- Rheology (1)
- Risikoanalyse (1)
- Risk analysis (1)
- Roth GmbH (1)
- RuBisCO (1)
- Russian-Jewish Elites (1)
- Russisch-jüdische Eliten (1)
- Räumlicher Wettbewerb (1)
- SHG (1)
- SOA (1)
- STG decomposition (1)
- STG-Dekomposition (1)
- Schaumfilme (1)
- Schemaentdeckung (1)
- Schwerefeld (1)
- Schätzung von Verzweigungsprozessen (1)
- Seasonality (1)
- Sedimentbecken (1)
- Selektion (1)
- Service Creation (1)
- Service Delivery Platform (1)
- Service convergence (1)
- Service-oriented Systems Engineering (1)
- Signalflankengraph (SFG oder STG) (1)
- Simulation (1)
- Smalltalk (1)
- Soliton (1)
- Spannungsfeld (1)
- Spermidin (1)
- Spin- und winkelaufgelöste Photoemission (1)
- Spinwellen (1)
- Spitzenbeamte (1)
- Spracherwerb (1)
- Sprachproduktion (1)
- Spurenelement-Partitionierung (1)
- Steuerlastverteilung (1)
- Steuerreform (1)
- Stimulus-Onset Asynchrony (1)
- Stock Prices (1)
- Story (1)
- Struktur (1)
- Studierendenaustausch (1)
- Styrol (1)
- Störungsinteraktion (1)
- Superoxid (1)
- Support-Vektor-Maschine (1)
- Supraleitender Gravimeter (SG) (1)
- Supramolecular Interaction (1)
- Supramolekularen Wechselwirkung (1)
- Synchrotronstrahlung (1)
- Synthese (1)
- System (1)
- Takab (1)
- Talmud (1)
- Tax distribution (1)
- Tectonic geomorphology (1)
- Tektonische Geomorphologie (1)
- Telekommunikation (1)
- Temporal Logic (1)
- Temporallogik (1)
- Theory of Mind (1)
- Tonoplast (1)
- Top-down (1)
- Topic (1)
- Transitional Justice (1)
- Transitions (1)
- Transnational Diaspora (1)
- Transnationale Diaspora (1)
- Transportkosten (1)
- Transtional Justice (1)
- Tsunami (1)
- Unscented Kalman Filter (1)
- Unterrichtseinheiten (1)
- VIL (1)
- VOC (1)
- Verben (1)
- Vereinigungs-Mapping (1)
- Verletzung Auflösung (1)
- Verletzung Erklärung (1)
- Verteilungsfunktionen mit einer oberen Grenze (1)
- Vertiefung (1)
- Violation Explanation (1)
- Violation Resolution (1)
- Vorhersagbarkeit (1)
- Vulcano (1)
- Vulkan (1)
- Vulkan Verformung (1)
- Wachstum (1)
- Wassermassenvariationen (1)
- Wasserspeicheränderungen (1)
- Web-Anwendungen (1)
- Wellengleichung (1)
- Wikipedia (1)
- Wühlmaus (1)
- XM (1)
- Zukunftsmärkte (1)
- Zwei-Länder-Modell (1)
- accelerated life time model (1)
- acyltransferase gene family (1)
- adaptive processes (1)
- affective priming (1)
- agrammatism (1)
- amino-acids (1)
- analysis (1)
- aquatic (1)
- aspect (1)
- aspect adapter (1)
- aspect oriented programming (1)
- aspect-oriented (1)
- aspects (1)
- aspectualization (1)
- association mapping (1)
- attribute analysis (1)
- automatic processes (1)
- bewaffnete Konflikte (1)
- bifurcation (1)
- bioactive (1)
- bioaktiv (1)
- biodiversity (1)
- biosensor (1)
- bistable regime (1)
- bottom water (1)
- bottom-up (1)
- bottom–up (1)
- bug tracking (1)
- built–in predicates (1)
- buraucratisation (1)
- cadmium tolerance (1)
- capital and ownership structure (1)
- cellular signalling (1)
- censoring (1)
- central Andes (1)
- centralization (1)
- chlorbenzol (1)
- climate change (1)
- climate warming (1)
- cobalt (1)
- community (1)
- complex networks (1)
- compliance (1)
- comprehension of complement sentences (1)
- concurrency (1)
- consistency (1)
- constitutive activity (1)
- control resynthesis (1)
- core executive (1)
- core field modeling (1)
- core flow modeling (1)
- coronal mass ejections (1)
- correlation (1)
- corruption (1)
- crosscutting wrappers (1)
- curriculum framework (1)
- cysteine synthase complex (1)
- cytochrome c (1)
- data integration (1)
- data profiling (1)
- de novo ncRNA Vorhersage (1)
- de novo ncRNA prediction (1)
- debugging (1)
- decentralization (1)
- deductive databases (1)
- degradation (1)
- dendritic network (1)
- dichlorbenzol (1)
- dichlorobenzene (1)
- dictyostelium cell (1)
- dietary lipids (1)
- dietary supplements (1)
- differential gene expression (1)
- direct taxes (1)
- distributed systems (1)
- double taxation (1)
- dynamic reconfiguration (1)
- earthquake (1)
- eclogite (UHP) (1)
- edge- and corner-degenerate symbols (1)
- efficient market hypothesis (1)
- electro-optical (1)
- electrochemistry (1)
- electrospinning (1)
- elektro-optisch (1)
- elite athletes (1)
- ellipticity (1)
- energy homeostasis (1)
- ereigniskorrelierte Potentiale (1)
- ethnic differences (1)
- eutrophication (1)
- evaluative priming (1)
- event-related potentials (1)
- evolving systems (1)
- exercise (1)
- exhumation rates (1)
- extensions of logic programs (1)
- eye tracking (1)
- ferroelectret (1)
- fiber (1)
- finite differences (1)
- fire (1)
- fiscal federalism (1)
- fiscal planning (1)
- fisheries (1)
- fluid-melt interaction (1)
- foam films (1)
- focus sensitive expressions (1)
- food web dynamics (1)
- forward / backward chaining (1)
- frequency (1)
- frozen-flux (1)
- function symbols (1)
- functional (1)
- functional-analysis (1)
- funktional (1)
- gelatin (1)
- gene expression matrix (1)
- generalized logic programs (1)
- genotyping (1)
- geochronology (1)
- geovirtual environments (1)
- geovirtuelle Umgebungen (1)
- global change (1)
- global sensitivity analysis (1)
- global water cycle (1)
- glutathione homeostasis (1)
- good governance (1)
- goodness-of-fit testing (1)
- granitoids (1)
- granulare Kette (1)
- graphene (1)
- gravity (1)
- ground penetrating radar (1)
- gydrological modelling (1)
- harmonization (1)
- health behavior (1)
- heterogeneous tissue (1)
- higher singularities (1)
- homogeneous cell population (1)
- human dynamics (1)
- hunger hormone (1)
- hydrogels (1)
- hydrological monitoring and modelling (1)
- hydrologische Modellierung (1)
- hydrology (1)
- höhere Singularitäten (1)
- ice dynamics (1)
- image processing (1)
- implicit (1)
- implizit (1)
- in-situ analysis (1)
- indirect taxes (1)
- individual based modeling (1)
- individuelle Modellierung (1)
- inflection (1)
- information exchange (1)
- information structure (1)
- insect (1)
- instrumental seismology (1)
- instrumentelle Seismologie (1)
- integration (1)
- inter- and intra-station variability (1)
- interfacial forces (1)
- international tax law (1)
- invasive aspects (1)
- inverse Modellierung (1)
- inverse modeling (1)
- iron (1)
- joint normal distribution (1)
- kanten- und ecken-entartete Symbole (1)
- knowledge representation (1)
- komplexe Netzwerke (1)
- konvergente Dienste (1)
- lake (1)
- language acquisition (1)
- language production (1)
- lesson evaluation (1)
- lexical access (1)
- linguistic determinism (1)
- linguistische Determinismushypothese (1)
- long memory (1)
- lower channel (1)
- magnetic field satellites (1)
- magnetic fields (1)
- magnetohydrodynamics (1)
- magnetometer calibration (1)
- management (1)
- map-based cloning (1)
- marine (1)
- menschliche Dynamik (1)
- mergers and acquisitions (1)
- metadata discovery (1)
- metadata quality (1)
- methodology (1)
- microdissection (1)
- microfluidic device (1)
- microsimulation (1)
- minerals (1)
- mitigation (1)
- model integration (1)
- model limitations (1)
- modelling (1)
- molecular modeling (1)
- monochlorobenzene (1)
- multi-stage exhumation (1)
- multi-valued logic (1)
- nanofluidics (1)
- nanoparticles (1)
- nation building (1)
- negation (1)
- nematic LC (1)
- non-linear dynamics (1)
- nonparametric regression estimation (1)
- numerical relativity (1)
- numerischen Relativitätstheorie (1)
- nutrients (1)
- o-acetylserine(thiol)lyase gene (1)
- ordinary differential equation (1)
- paraconsistency (1)
- parafoveal preview (1)
- parametrices (1)
- partial melting (1)
- persistence (1)
- petri net (1)
- phenology (1)
- phosphorylated amino acid (1)
- photoelectron microscopy (1)
- phytoplankton (1)
- piezo-optical (1)
- piezo-optisch (1)
- piezoelectricity (1)
- plankton (1)
- plant functional types (1)
- politicisation (1)
- polymer film (1)
- polymer foam (1)
- polymer networks (1)
- polynya (1)
- population (1)
- porous scaffolds (1)
- poröse Gerüste (1)
- predictability (1)
- prediction (1)
- profiling (1)
- program analysis (1)
- programming (1)
- programming language (1)
- protein engineering (1)
- proteome (1)
- psycholinguistics (1)
- qPCR (1)
- quantile normalization (1)
- raumbezogene Straftatenanalyse (1)
- reflection (1)
- regenerative Medizin (1)
- regenerative medicine (1)
- rekonfigurierbar (1)
- reusable aspects (1)
- revenue authorities (1)
- revenue sharing (1)
- schema discovery (1)
- security policies (1)
- sedimentary basin (1)
- seismische Rauschen (1)
- seismisches Rauschen (1)
- selection (1)
- semantic fieldwork (1)
- semantic priming (1)
- semantic web services (1)
- semantics (1)
- sentence processing (1)
- serine acetyltransferase (1)
- service mediation (1)
- service-oriented (1)
- signal transition graph (1)
- simulation (1)
- single nucleotide polymorphism mapping (1)
- slim term (1)
- smalltalk (1)
- source parameters (1)
- spatial (1)
- species richness (1)
- speed independent (1)
- spermidine (1)
- spin waves (1)
- spin- and angle-resolved photoemission (1)
- stable fixed point (1)
- stem selection (1)
- stimulus-onset asynchrony (1)
- stock market (1)
- stock returns (1)
- stratification (1)
- stress field (1)
- structure (1)
- student exchange (1)
- styrene (1)
- sulfur metabolism (1)
- sun (1)
- superconducting gravimeter (SG) (1)
- superoxide (1)
- support vector machine (1)
- sustainable fiscal policy (1)
- synchrotron radiation (1)
- syntax (1)
- tax administration (1)
- tax harmonization (1)
- tax policy evaluation (1)
- tax system (1)
- taxation (1)
- teaching units (1)
- temporal gravity variations (1)
- theory of mind (1)
- tolerable upper limits (1)
- top bureaucrats (1)
- top-down (1)
- top– down (1)
- total internal reflection fluorescence (1)
- trace element partitioning (1)
- transparency (1)
- tuple spaces (1)
- understanding (1)
- verbs (1)
- views (1)
- visual word recognition (1)
- visuelle Worterkennung (1)
- vitamins (1)
- volcano deformation (1)
- water mass variations (1)
- water storage changes (1)
- wave equation (1)
- web services (1)
- web-applications (1)
- word processing (1)
- work ethics (1)
- zeitabhängige Gravitationsvariation (1)
- zentralen Anden (1)
- Übergangsprozesse (1)
Institute
- Extern (22)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (17)
- Institut für Geowissenschaften (16)
- Institut für Jüdische Studien und Religionswissenschaft (16)
- Vereinigung für Jüdische Studien e. V. (15)
- Institut für Biochemie und Biologie (13)
- Institut für Künste und Medien (11)
- Mathematisch-Naturwissenschaftliche Fakultät (11)
- Institut für Mathematik (9)
- Wirtschaftswissenschaften (9)
Situated in an active tectonic region, Santiago de Chile, the country´s capital with more than six million inhabitants, faces tremendous earthquake hazard. Macroseismic data for the 1985 Valparaiso and the 2010 Maule events show large variations in the distribution of damage to buildings within short distances indicating strong influence of local sediments and the shape of the sediment-bedrock interface on ground motion. Therefore, a temporary seismic network was installed in the urban area for recording earthquake activity, and a study was carried out aiming to estimate site amplification derived from earthquake data and ambient noise. The analysis of earthquake data shows significant dependence on the local geological structure with regards to amplitude and duration. Moreover, the analysis of noise spectral ratios shows that they can provide a lower bound in amplitude for site amplification and, since no variability in terms of time and amplitude is observed, that it is possible to map the fundamental resonance frequency of the soil for a 26 km x 12 km area in the northern part of the Santiago de Chile basin. By inverting the noise spectral rations, local shear wave velocity profiles could be derived under the constraint of the thickness of the sedimentary cover which had previously been determined by gravimetric measurements. The resulting 3D model was derived by interpolation between the single shear wave velocity profiles and shows locally good agreement with the few existing velocity profile data, but allows the entire area, as well as deeper parts of the basin, to be represented in greater detail. The wealth of available data allowed further to check if any correlation between the shear wave velocity in the uppermost 30 m (vs30) and the slope of topography, a new technique recently proposed by Wald and Allen (2007), exists on a local scale. While one lithology might provide a greater scatter in the velocity values for the investigated area, almost no correlation between topographic gradient and calculated vs30 exists, whereas a better link is found between vs30 and the local geology. When comparing the vs30 distribution with the MSK intensities for the 1985 Valparaiso event it becomes clear that high intensities are found where the expected vs30 values are low and over a thick sedimentary cover. Although this evidence cannot be generalized for all possible earthquakes, it indicates the influence of site effects modifying the ground motion when earthquakes occur well outside of the Santiago basin. Using the attained knowledge on the basin characteristics, simulations of strong ground motion within the Santiago Metropolitan area were carried out by means of the spectral element technique. The simulation of a regional event, which has also been recorded by a dense network installed in the city of Santiago for recording aftershock activity following the 27 February 2010 Maule earthquake, shows that the model is capable to realistically calculate ground motion in terms of amplitude, duration, and frequency and, moreover, that the surface topography and the shape of the sediment bedrock interface strongly modify ground motion in the Santiago basin. An examination on the dependency of ground motion on the hypocenter location for a hypothetical event occurring along the active San Ramón fault, which is crossing the eastern outskirts of the city, shows that the unfavorable interaction between fault rupture, radiation mechanism, and complex geological conditions in the near-field may give rise to large values of peak ground velocity and therefore considerably increase the level of seismic risk for Santiago de Chile.
Companies develop process models to explicitly describe their business operations. In the same time, business operations, business processes, must adhere to various types of compliance requirements. Regulations, e.g., Sarbanes Oxley Act of 2002, internal policies, best practices are just a few sources of compliance requirements. In some cases, non-adherence to compliance requirements makes the organization subject to legal punishment. In other cases, non-adherence to compliance leads to loss of competitive advantage and thus loss of market share. Unlike the classical domain-independent behavioral correctness of business processes, compliance requirements are domain-specific. Moreover, compliance requirements change over time. New requirements might appear due to change in laws and adoption of new policies. Compliance requirements are offered or enforced by different entities that have different objectives behind these requirements. Finally, compliance requirements might affect different aspects of business processes, e.g., control flow and data flow. As a result, it is infeasible to hard-code compliance checks in tools. Rather, a repeatable process of modeling compliance rules and checking them against business processes automatically is needed. This thesis provides a formal approach to support process design-time compliance checking. Using visual patterns, it is possible to model compliance requirements concerning control flow, data flow and conditional flow rules. Each pattern is mapped into a temporal logic formula. The thesis addresses the problem of consistency checking among various compliance requirements, as they might stem from divergent sources. Also, the thesis contributes to automatically check compliance requirements against process models using model checking. We show that extra domain knowledge, other than expressed in compliance rules, is needed to reach correct decisions. In case of violations, we are able to provide a useful feedback to the user. The feedback is in the form of parts of the process model whose execution causes the violation. In some cases, our approach is capable of providing automated remedy of the violation.
Background: Leishmania tarentolae, a unicellular eukaryotic protozoan, has been established as a novel host for recombinant protein production in recent years. Current protocols for protein expression in Leishmania are, however, time consuming and require extensive lab work in order to identify well-expressing cell lines. Here we established an alternative protein expression work-flow that employs recently engineered infrared fluorescence protein (IFP) as a suitable and easy-to-handle reporter protein for recombinant protein expression in Leishmania. As model proteins we tested three proteins from the plant Arabidopsis thaliana, including a NAC and a type-B ARR transcription factor. Results: IFP and IFP fusion proteins were expressed in Leishmania and rapidly detected in cells by deconvolution microscopy and in culture by infrared imaging of 96-well microtiter plates using small cell culture volumes (2 μL - 100 μL). Motility, shape and growth of Leishmania cells were not impaired by intracellular accumulation of IFP. In-cell detection of IFP and IFP fusion proteins was straightforward already at the beginning of the expression pipeline and thus allowed early pre-selection of well-expressing Leishmania clones. Furthermore, IFP fusion proteins retained infrared fluorescence after electrophoresis in denaturing SDS-polyacrylamide gels, allowing direct in-gel detection without the need to disassemble cast protein gels. Thus, parameters for scaling up protein production and streamlining purification routes can be easily optimized when employing IFP as reporter. Conclusions: Using IFP as biosensor we devised a protocol for rapid and convenient protein expression in Leishmania tarentolae. Our expression pipeline is superior to previously established methods in that it significantly reduces the hands-on-time and work load required for identifying well-expressing clones, refining protein production parameters and establishing purification protocols. The facile in-cell and in-gel detection tools built on IFP make Leishmania amenable for high-throughput expression of proteins from plant and animal sources.
Background: Cysteine is a component in organic compounds including glutathione that have been implicated in the adaptation of plants to stresses. O-acetylserine (thiol) lyase (OAS-TL) catalyses the final step of cysteine biosynthesis. OAS-TL enzyme isoforms are localised in the cytoplasm, the plastids and mitochondria but the contribution of individual OAS-TL isoforms to plant sulphur metabolism has not yet been fully clarified.
Results: The seedling lethal phenotype of the Arabidopsis onset of leaf death3-1 (old3-1) mutant is due to a point mutation in the OAS-A1 gene, encoding the cytosolic OAS-TL. The mutation causes a single amino acid substitution from Gly(162) to Glu(162), abolishing old3-1 OAS-TL activity in vitro. The old3-1 mutation segregates as a monogenic semidominant trait when backcrossed to its wild type accession Landsberg erecta (Ler-0) and the Di-2 accession. Consistent with its semi-dominant behaviour, wild type Ler-0 plants transformed with the mutated old3-1 gene, displayed the early leaf death phenotype. However, the old3-1 mutation segregates in an 11: 4: 1 (wild type: semi-dominant: mutant) ratio when backcrossed to the Colombia-0 and Wassilewskija accessions. Thus, the early leaf death phenotype depends on two semi-dominant loci. The second locus that determines the old3-1 early leaf death phenotype is referred to as odd-ler (for old3 determinant in the Ler accession) and is located on chromosome 3. The early leaf death phenotype is temperature dependent and is associated with increased expression of defence-response and oxidative-stress marker genes. Independent of the presence of the odd-ler gene, OAS-A1 is involved in maintaining sulphur and thiol levels and is required for resistance against cadmium stress.
Conclusions: The cytosolic OAS-TL is involved in maintaining organic sulphur levels. The old3-1 mutation causes genome-dependent and independent phenotypes and uncovers a novel function for the mutated OAS-TL in cell death regulation.
We propose a paraconsistent declarative semantics of possibly inconsistent generalized logic programs which allows for arbitrary formulas in the body and in the head of a rule (i.e. does not depend on the presence of any specific connective, such as negation(-as-failure), nor on any specific syntax of rules). For consistent generalized logic programs this semantics coincides with the stable generated models introduced in [HW97], and for normal logic programs it yields the stable models in the sense of [GL88].
We reconsider the fundamental work of Fichtner ([2]) and exhibit the permanental structure of the ideal Bose gas again, using another approach which combines a characterization of infinitely divisible random measures (due to Kerstan,Kummer and Matthes [5, 6] and Mecke [8, 9]) with a decomposition of the moment measures into its factorial measures due to Krickeberg [4]. To be more precise, we exhibit the moment measures of all orders of the general ideal Bose gas in terms of certain path integrals. This representation can be considered as a point process analogue of the old idea of Symanzik [11] that local times and self-crossings of the Brownian motion can be used as a tool in quantum field theory. Behind the notion of a general ideal Bose gas there is a class of infinitely divisible point processes of all orders with a Levy-measure belonging to some large class of measures containing the one of the classical ideal Bose gas considered by Fichtner. It is well known that the calculation of moments of higher order of point processes are notoriously complicated. See for instance Krickeberg's calculations for the Poisson or the Cox process in [4].
This thesis is focused on the electronic, spin-dependent and dynamical properties of thin magnetic systems. Photoemission-related techniques are combined with synchrotron radiation to study the spin-dependent properties of these systems in the energy and time domains. In the first part of this thesis, the strength of electron correlation effects in the spin-dependent electronic structure of ferromagnetic bcc Fe(110) and hcp Co(0001) is investigated by means of spin- and angle-resolved photoemission spectroscopy. The experimental results are compared to theoretical calculations within the three-body scattering approximation and within the dynamical mean-field theory, together with one-step model calculations of the photoemission process. From this comparison it is demonstrated that the present state of the art many-body calculations, although improving the description of correlation effects in Fe and Co, give too small mass renormalizations and scattering rates thus demanding more refined many-body theories including nonlocal fluctuations. In the second part, it is shown in detail monitoring by photoelectron spectroscopy how graphene can be grown by chemical vapour deposition on the transition-metal surfaces Ni(111) and Co(0001) and intercalated by a monoatomic layer of Au. For both systems, a linear E(k) dispersion of massless Dirac fermions is observed in the graphene pi-band in the vicinity of the Fermi energy. Spin-resolved photoemission from the graphene pi-band shows that the ferromagnetic polarization of graphene/Ni(111) and graphene/Co(0001) is negligible and that graphene on Ni(111) is after intercalation of Au spin-orbit split by the Rashba effect. In the last part, a time-resolved x-ray magnetic circular dichroic-photoelectron emission microscopy study of a permalloy platelet comprising three cross-tie domain walls is presented. It is shown how a fast picosecond magnetic response in the precessional motion of the magnetization can be induced by means of a laser-excited photoswitch. From a comparision to micromagnetic calculations it is demonstrated that the relatively high precessional frequency observed in the experiments is directly linked to the nature of the vortex/antivortex dynamics and its response to the magnetic perturbation. This includes the time-dependent reversal of the vortex core polarization, a process which is beyond the limit of detection in the present experiments.
A deterministic cycle scheduling of partitions at the operating system level is supposed for a multiprocessor system. In this paper, we propose a tool for generating such schedules. We use constraint based programming and develop methods and concepts for a combined interactive and automatic partition scheduling system. This paper is also devoted to basic methods and techniques for modeling and solving this partition scheduling problem. Initial application of our partition scheduling tool has proved successful and demonstrated the suitability of the methods used.
About the relation between implicit Theory of Mind & the comprehension of complement sentences
(2010)
Previous studies on the relation between language and social cognition have shown that children’s mastery of embedded sentential complements plays a causal role for the development of a Theory of Mind (ToM). Children start to succeed on complementation tasks in which they are required to report the content of an embedded clause in the second half of the fourth year. Traditional ToM tasks test the child’s ability to predict that a person who is holding a false belief (FB) about a situation will act "falsely". In these task, children do not represent FBs until the age of 4 years. According the linguistic determinism hypothesis, only the unique syntax of complement sentences provides the format for representing FBs. However, experiments measuring children’s looking behavior instead of their explicit predictions provided evidence that already 2-year olds possess an implicit ToM. This dissertation examined the question of whether there is an interrelation also between implicit ToM and the comprehension of complement sentences in typically developing German preschoolers. Two studies were conducted. In a correlational study (Study 1 ), 3-year-old children’s performance on a traditional (explicit) FB task, on an implicit FB task and on language tasks measuring children’s comprehension of tensed sentential complements were collected and tested for their interdependence. Eye-tracking methodology was used to assess implicit ToM by measuring participants’ spontaneous anticipatory eye movements while they were watching FB movies. Two central findings emerged. First, predictive looking (implicit ToM) was not correlated with complement mastery, although both measures were associated with explicit FB task performance. This pattern of results suggests that explicit, but not implicit ToM is language dependent. Second, as a group, 3-year-olds did not display implicit FB understanding. That is, previous findings on a precocious reasoning ability could not be replicated. This indicates that the characteristics of predictive looking tasks play a role for the elicitation of implicit FB understanding as the current task was completely nonverbal and as complex as traditional FB tasks. Study 2 took a methodological approach by investigating whether children display an earlier comprehension of sentential complements when using the same means of measurement as used in experimental tasks tapping implicit ToM, namely anticipatory looking. Two experiments were conducted. 3-year-olds were confronted either with a complement sentence expressing the protagonist’s FB (Exp. 1) or with a complex sentence expressing the protagonist’s belief without giving any information about the truth/ falsity of the belief (Exp. 2). Afterwards, their expectations about the protagonist’s future behavior were measured. Overall, implicit measures reveal no considerably earlier understanding of sentential complementation. Whereas 3-year-olds did not display a comprehension of complex sentences if these embedded a false proposition, children from 3;9 years on were proficient in processing complement sentences if the truth value of the embedded proposition could not be evaluated. This pattern of results suggests that (1) the linguistic expression of a person’s FB does not elicit implicit FB understanding and that (2) the assessment of the purely syntactic understanding of complement sentences is affected by competing reality information. In conclusion, this dissertation found no evidence that the implicit ToM is related to the comprehension of sentential complementation. The findings suggest that implicit ToM might be based on nonlinguistic processes. Results are discussed in the light of recently proposed dual-process models that assume two cognitive mechanisms that account for different levels of ToM task performance.
We describe a framework to support the implementation of web-based systems to manipulate data stored in relational databases. Since the conceptual model of a relational database is often specified as an entity-relationship (ER) model, we propose to use the ER model to generate a complete implementation in the declarative programming language Curry. This implementation contains operations to create and manipulate entities of the data model, supports authentication, authorization, session handling, and the composition of individual operations to user processes. Furthermore and most important, the implementation ensures the consistency of the database w.r.t. the data dependencies specified in the ER model, i.e., updates initiated by the user cannot lead to an inconsistent state of the database. In order to generate a high-level declarative implementation that can be easily adapted to individual customer requirements, the framework exploits previous works on declarative database programming and web user interface construction in Curry.
We establish elements of a new approach to ellipticity and parametrices within operator algebras on manifolds with higher singularities, only based on some general axiomatic requirements on parameter-dependent operators in suitable scales of spaes. The idea is to model an iterative process with new generations of parameter-dependent operator theories, together with new scales of spaces that satisfy analogous requirements as the original ones, now on a corresponding higher level. The "full" calculus involves two separate theories, one near the tip of the corner and another one at the conical exit to infinity. However, concerning the conical exit to infinity, we establish here a new concrete calculus of edge-degenerate operators which can be iterated to higher singularities.
Persistence of stock returns is an extensively studied and discussed theme in the analysis of financial markets. Antipersistence is usually attributed to volatilities. However, not only volatilities but also stock returns can exhibit antipersistence. Antipersistent noise has a somewhat rougher appearance than Gaussian noise. Heuristically spoken, price movements are more likely followed by movements in the opposite direction than in the same direction. The pertaining integrated process exhibits a smaller range – prices seem to stay in the vicinity of the initial value. We apply a widely used test based upon the modified R/S-Method by Lo [1991] to daily returns of 21 German stocks from 1960 to 2008. Combining this test with the concept of moving windows by Carbone et al. [2004], we are able to determine periods of antipersistence for some of the series under examination. Our results suggest that antipersistence can be found for stocks and periods where extraordinary corporate actions such as mergers & acquisitions or financial distress are present. These effects should be properly accounted for when choosing and designing models for inference.
This thesis presents methods for automated synthesis of flexible chip multiprocessor systems from parallel programs targeted at FPGAs to exploit both task-level parallelism and architecture customization. Automated synthesis is necessitated by the complexity of the design space. A detailed description of the design space is provided in order to determine which parameters should be modeled to facilitate automated synthesis by optimizing a cost function, the emphasis being placed on inclusive modeling of parameters from application, architectural and physical subspaces, as well as their joint coverage in order to avoid pre-constraining the design space. Given a parallel program and a set of an IP library, the automated synthesis problem is to simultaneously (i) select processors (ii) map and schedule tasks to them, and (iii) select one or several networks for inter-task communications such that design constraints and optimization objectives are met. The research objective in this thesis is to find a suitable model for automated synthesis, and to evaluate methods of using the model for architectural optimizations. Our contributions are a holistic approach for the design of such systems, corresponding models to facilitate automated synthesis, evaluation of optimization methods using state of the art integer linear and answer set programming, as well as the development of synthesis heuristics to solve runtime challenges.
Enforcing security policies to distributed systems is difficult, in particular, when a system contains untrusted components. We designed AspectKE*, a distributed AOP language based on a tuple space, to tackle this issue. In AspectKE*, aspects can enforce access control policies that depend on future behavior of running processes. One of the key language features is the predicates and functions that extract results of static program analysis, which are useful for defining security aspects that have to know about future behavior of a program. AspectKE* also provides a novel variable binding mechanism for pointcuts, so that pointcuts can uniformly specify join points based on both static and dynamic information about the program. Our implementation strategy performs fundamental static analysis at load-time, so as to retain runtime overheads minimal. We implemented a compiler for AspectKE*, and demonstrate usefulness of AspectKE* through a security aspect for a distributed chat system.
Based on technological advances made within the past decades, ground-penetrating radar (GPR) has become a well-established, non-destructive subsurface imaging technique. Catalyzed by recent demands for high-resolution, near-surface imaging (e.g., the detection of unexploded ordnances and subsurface utilities, or hydrological investigations), the quality of today's GPR-based, near-surface images has significantly matured. At the same time, the analysis of oil and gas related reflection seismic data sets has experienced significant advances. Considering the sensitivity of attribute analysis with respect to data positioning in general, and multi-trace attributes in particular, trace positioning accuracy is of major importance for the success of attribute-based analysis flows. Therefore, to study the feasibility of GPR-based attribute analyses, I first developed and evaluated a real-time GPR surveying setup based on a modern tracking total station (TTS). The combination of current GPR systems capability of fusing global positioning system (GPS) and geophysical data in real-time, the ability of modern TTS systems to generate a GPS-like positional output and wireless data transmission using radio modems results in a flexible and robust surveying setup. To elaborate the feasibility of this setup, I studied the major limitations of such an approach: system cross-talk and data delays known as latencies. Experimental studies have shown that when a minimal distance of ~5 m between the GPR and the TTS system is considered, the signal-to-noise ratio of the acquired GPR data using radio communication equals the one without radio communication. To address the limitations imposed by system latencies, inherent to all real-time data fusion approaches, I developed a novel correction (calibration) strategy to assess the gross system latency and to correct for it. This resulted in the centimeter trace accuracy required by high-frequency and/or three-dimensional (3D) GPR surveys. Having introduced this flexible high-precision surveying setup, I successfully demonstrated the application of attribute-based processing to GPR specific problems, which may differ significantly from the geological ones typically addressed by the oil and gas industry using seismic data. In this thesis, I concentrated on archaeological and subsurface utility problems, as they represent typical near-surface geophysical targets. Enhancing 3D archaeological GPR data sets using a dip-steered filtering approach, followed by calculation of coherency and similarity, allowed me to conduct subsurface interpretations far beyond those obtained by classical time-slice analyses. I could show that the incorporation of additional data sets (magnetic and topographic) and attributes derived from these data sets can further improve the interpretation. In a case study, such an approach revealed the complementary nature of the individual data sets and, for example, allowed conclusions about the source location of magnetic anomalies by concurrently analyzing GPR time/depth slices to be made. In addition to archaeological targets, subsurface utility detection and characterization is a steadily growing field of application for GPR. I developed a novel attribute called depolarization. Incorporation of geometrical and physical feature characteristics into the depolarization attribute allowed me to display the observed polarization phenomena efficiently. Geometrical enhancement makes use of an improved symmetry extraction algorithm based on Laplacian high-boosting, followed by a phase-based symmetry calculation using a two-dimensional (2D) log-Gabor filterbank decomposition of the data volume. To extract the physical information from the dual-component data set, I employed a sliding-window principle component analysis. The combination of the geometrically derived feature angle and the physically derived polarization angle allowed me to enhance the polarization characteristics of subsurface features. Ground-truth information obtained by excavations confirmed this interpretation. In the future, inclusion of cross-polarized antennae configurations into the processing scheme may further improve the quality of the depolarization attribute. In addition to polarization phenomena, the time-dependent frequency evolution of GPR signals might hold further information on the subsurface architecture and/or material properties. High-resolution, sparsity promoting decomposition approaches have recently had a significant impact on the image and signal processing community. In this thesis, I introduced a modified tree-based matching pursuit approach. Based on different synthetic examples, I showed that the modified tree-based pursuit approach clearly outperforms other commonly used time-frequency decomposition approaches with respect to both time and frequency resolutions. Apart from the investigation of tuning effects in GPR data, I also demonstrated the potential of high-resolution sparse decompositions for advanced data processing. Frequency modulation of individual atoms themselves allows to efficiently correct frequency attenuation effects and improve resolution based on shifting the average frequency level. GPR-based attribute analysis is still in its infancy. Considering the growing widespread realization of 3D GPR studies there will certainly be an increasing demand towards improved subsurface interpretations in the future. Similar to the assessment of quantitative reservoir properties through the combination of 3D seismic attribute volumes with sparse well-log information, parameter estimation in a combined manner represents another step in emphasizing the potential of attribute-driven GPR data analyses.
Within our research group Bayesian Risk Solutions we have coined the idea of a Bayesian Risk Management (BRM). It claims (1) a more transparent and diligent data analysis as well as (2)an open-minded incorporation of human expertise in risk management. In this dissertation we formulize a framework for BRM based on the two pillars Hardcore-Bayesianism (HCB) and Softcore-Bayesianism (SCB) providing solutions for the claims above. For data analysis we favor Bayesian statistics with its Markov Chain Monte Carlo (MCMC) simulation algorithm. It provides a full illustration of data-induced uncertainty beyond classical point-estimates. We calibrate twelve different stochastic processes to four years of CO2 price data. Besides, we calculate derived risk measures (ex ante/ post value-at-risks, capital charges, option prices) and compare them to their classical counterparts. When statistics fails because of a lack of reliable data we propose our integrated Bayesian Risk Analysis (iBRA) concept. It is a basic guideline for an expertise-driven quantification of critical risks. We additionally review elicitation techniques and tools supporting experts to express their uncertainty. Unfortunately, Bayesian thinking is often blamed for its arbitrariness. Therefore, we introduce the idea of a Bayesian due diligence judging expert assessments according to their information content and their inter-subjectivity.
Between history and legend
(2010)
In the early modern period, Jewish historiography moved from the Hebrew domain into the Yiddish one. Jewish writers have succeeded to match the historical literature to the particular needs of their audience. The most popular Yiddish chronicle of this kind was written in Amsterdam in the 18th century by Menachem Man Amelander, following both the Jewish and Christian genre. This paper briefly surveys the genre characteristics of this chronicle and the way it served the purpose of guarding Jewish memory and tradition.
The genome can be considered the blueprint for an organism. Composed of DNA, it harbours all organism-specific instructions for the synthesis of all structural components and their associated functions. The role of carriers of actual molecular structure and functions was believed to be exclusively assumed by proteins encoded in particular segments of the genome, the genes. In the process of converting the information stored genes into functional proteins, RNA – a third major molecule class – was discovered early on to act a messenger by copying the genomic information and relaying it to the protein-synthesizing machinery. Furthermore, RNA molecules were identified to assist in the assembly of amino acids into native proteins. For a long time, these - rather passive - roles were thought to be the sole purpose of RNA. However, in recent years, new discoveries have led to a radical revision of this view. First, RNA molecules with catalytic functions - thought to be the exclusive domain of proteins - were discovered. Then, scientists realized that much more of the genomic sequence is transcribed into RNA molecules than there are proteins in cells begging the question what the function of all these molecules are. Furthermore, very short and altogether new types of RNA molecules seemingly playing a critical role in orchestrating cellular processes were discovered. Thus, RNA has become a central research topic in molecular biology, even to the extent that some researcher dub cells as “RNA machines”. This thesis aims to contribute towards our understanding of RNA-related phenomena by applying Bioinformatics means. First, we performed a genome-wide screen to identify sites at which the chemical composition of DNA (the genotype) critically influences phenotypic traits (the phenotype) of the model plant Arabidopsis thaliana. Whole genome hybridisation arrays were used and an informatics strategy developed, to identify polymorphic sites from hybridisation to genomic DNA. Following this approach, not only were genotype-phenotype associations discovered across the entire Arabidopsis genome, but also regions not currently known to encode proteins, thus representing candidate sites for novel RNA functional molecules. By statistically associating them with phenotypic traits, clues as to their particular functions were obtained. Furthermore, these candidate regions were subjected to a novel RNA-function classification prediction method developed as part of this thesis. While determining the chemical structure (the sequence) of candidate RNA molecules is relatively straightforward, the elucidation of its structure-function relationship is much more challenging. Towards this end, we devised and implemented a novel algorithmic approach to predict the structural and, thereby, functional class of RNA molecules. In this algorithm, the concept of treating RNA molecule structures as graphs was introduced. We demonstrate that this abstraction of the actual structure leads to meaningful results that may greatly assist in the characterization of novel RNA molecules. Furthermore, by using graph-theoretic properties as descriptors of structure, we indentified particular structural features of RNA molecules that may determine their function, thus providing new insights into the structure-function relationships of RNA. The method (termed Grapple) has been made available to the scientific community as a web-based service. RNA has taken centre stage in molecular biology research and novel discoveries can be expected to further solidify the central role of RNA in the origin and support of life on earth. As illustrated by this thesis, Bioinformatics methods will continue to play an essential role in these discoveries.
Background: For heterogeneous tissues, such as blood, measurements of gene expression are confounded by relative proportions of cell types involved. Conclusions have to rely on estimation of gene expression signals for homogeneous cell populations, e.g. by applying micro-dissection, fluorescence activated cell sorting, or in-silico deconfounding. We studied feasibility and validity of a non-negative matrix decomposition algorithm using experimental gene expression data for blood and sorted cells from the same donor samples. Our objective was to optimize the algorithm regarding detection of differentially expressed genes and to enable its use for classification in the difficult scenario of reversely regulated genes. This would be of importance for the identification of candidate biomarkers in heterogeneous tissues.
Results: Experimental data and simulation studies involving noise parameters estimated from these data revealed that for valid detection of differential gene expression, quantile normalization and use of non-log data are optimal. We demonstrate the feasibility of predicting proportions of constituting cell types from gene expression data of single samples, as a prerequisite for a deconfounding-based classification approach. Classification cross-validation errors with and without using deconfounding results are reported as well as sample-size dependencies. Implementation of the algorithm, simulation and analysis scripts are available.
Conclusions: The deconfounding algorithm without decorrelation using quantile normalization on non-log data is proposed for biomarkers that are difficult to detect, and for cases where confounding by varying proportions of cell types is the suspected reason. In this case, a deconfounding ranking approach can be used as a powerful alternative to, or complement of, other statistical learning approaches to define candidate biomarkers for molecular diagnosis and prediction in biomedicine, in realistically noisy conditions and with moderate sample sizes.
Multi-color fluorescence imaging experiments of wave forming Dictyostelium cells have revealed that actin waves separate two domains of the cell cortex that differ in their actin structure and phosphoinositide composition. We propose a bistable model of actin dynamics to account for these experimental observation. The model is based on the simplifying assumption that the actin cytoskeleton is composed of two distinct network types, a dendritic and a bundled network. The two structurally different states that were observed in experiments correspond to the stable fixed points in the bistable regime of this model. Each fixed point is dominated by one of the two network types. The experimentally observed actin waves can be considered as trigger waves that propagate transitions between the two stable fixed points.
In reading, word frequency is commonly regarded as the major bottom-up determinant for the speed of lexical access. Moreover, language processing depends on top-down information, such as the predictability of a word from a previous context. Yet, however, the exact role of top-down predictions in visual word recognition is poorly understood: They may rapidly affect lexical processes, or alternatively, influence only late post-lexical stages. To add evidence about the nature of top-down processes and their relation to bottom-up information in the timeline of word recognition, we examined influences of frequency and predictability on event-related potentials (ERPs) in several sentence reading studies. The results were related to eye movements from natural reading as well as to models of word recognition. As a first and major finding, interactions of frequency and predictability on ERP amplitudes consistently revealed top-down influences on lexical levels of word processing (Chapters 2 and 4). Second, frequency and predictability mediated relations between N400 amplitudes and fixation durations, pointing to their sensitivity to a common stage of word recognition; further, larger N400 amplitudes entailed longer fixation durations on the next word, a result providing evidence for ongoing processing beyond a fixation (Chapter 3). Third, influences of presentation rate on ERP frequency and predictability effects demonstrated that the time available for word processing critically co-determines the course of bottom-up and top-down influences (Chapter 4). Fourth, at a near-normal reading speed, an early predictability effect suggested the rapid comparison of top-down hypotheses with the actual visual input (Chapter 5). The present results are compatible with interactive models of word recognition assuming that early lexical processes depend on the concerted impact of bottom-up and top-down information. We offered a framework that reconciles the findings on a timeline of word recognition taking into account influences of frequency, predictability, and presentation rate (Chapter 4).
Business process management aims at capturing, understanding, and improving work in organizations. The central artifacts are process models, which serve different purposes. Detailed process models are used to analyze concrete working procedures, while high-level models show, for instance, handovers between departments. To provide different views on process models, business process model abstraction has emerged. While several approaches have been proposed, a number of abstraction use case that are both relevant for industry and scientifically challenging are yet to be addressed. In this paper we systematically develop, classify, and consolidate different use cases for business process model abstraction. The reported work is based on a study with BPM users in the health insurance sector and validated with a BPM consultancy company and a large BPM vendor. The identified fifteen abstraction use cases reflect the industry demand. The related work on business process model abstraction is evaluated against the use cases, which leads to a research agenda.
Calibration of the global hydrological model WGHM with water mass variations from GRACE gravity data
(2010)
Since the start-up of the GRACE (Gravity Recovery And Climate Experiment) mission in 2002 time dependent global maps of the Earth's gravity field are available to study geophysical and climatologically-driven mass redistributions on the Earth's surface. In particular, GRACE observations of total water storage changes (TWSV) provide a comprehensive data set for analysing the water cycle on large scales. Therefore they are invaluable for validation and calibration of large-scale hydrological models as the WaterGAP Global Hydrology Model (WGHM) which simulates the continental water cycle including its most important components, such as soil, snow, canopy, surface- and groundwater. Hitherto, WGHM exhibits significant differences to GRACE, especially for the seasonal amplitude of TWSV. The need for a validation of hydrological models is further highlighted by large differences between several global models, e.g. WGHM, the Global Land Data Assimilation System (GLDAS) and the Land Dynamics model (LaD). For this purpose, GRACE links geodetic and hydrological research aspects. This link demands the development of adequate data integration methods on both sides, forming the main objectives of this work. They include the derivation of accurate GRACE-based water storage changes, the development of strategies to integrate GRACE data into a global hydrological model as well as a calibration method, followed by the re-calibration of WGHM in order to analyse process and model responses. To achieve these aims, GRACE filter tools for the derivation of regionally averaged TWSV were evaluated for specific river basins. Here, a decorrelation filter using GRACE orbits for its design is most efficient among the tested methods. Consistency in data and equal spatial resolution between observed and simulated TWSV were realised by the inclusion of all most important hydrological processes and an equal filtering of both data sets. Appropriate calibration parameters were derived by a WGHM sensitivity analysis against TWSV. Finally, a multi-objective calibration framework was developed to constrain model predictions by both river discharge and GRACE TWSV, realised with a respective evolutionary method, the ε-Non-dominated-Sorting-Genetic-Algorithm-II (ε-NSGAII). Model calibration was done for the 28 largest river basins worldwide and for most of them improved simulation results were achieved with regard to both objectives. From the multi-objective approach more reliable and consistent simulations of TWSV within the continental water cycle were gained and possible model structure errors or mis-modelled processes for specific river basins detected. For tropical regions as such, the seasonal amplitude of water mass variations has increased. The findings lead to an improved understanding of hydrological processes and their representation in the global model. Finally, the robustness of the results is analysed with respect to GRACE and runoff measurement errors. As a main conclusion obtained from the results, not only soil water and snow storage but also groundwater and surface water storage have to be included in the comparison of the modelled and GRACE-derived total water budged data. Regarding model calibration, the regional varying distribution of parameter sensitivity suggests to tune only parameter of important processes within each region. Furthermore, observations of single storage components beside runoff are necessary to improve signal amplitudes and timing of simulated TWSV as well as to evaluate them with higher accuracy. The results of this work highlight the valuable nature of GRACE data when merged into large-scale hydrological modelling and depict methods to improve large-scale hydrological models.
A large number and wide variety of lake ecosystem models have been developed and published during the past four decades. We identify two challenges for making further progress in this field. One such challenge is to avoid developing more models largely following the concept of others ('reinventing the wheel'). The other challenge is to avoid focusing on only one type of model, while ignoring new and diverse approaches that have become available ('having tunnel vision'). In this paper, we aim at improving the awareness of existing models and knowledge of concurrent approaches in lake ecosystem modelling, without covering all possible model tools and avenues. First, we present a broad variety of modelling approaches. To illustrate these approaches, we give brief descriptions of rather arbitrarily selected sets of specific models. We deal with static models (steady state and regression models), complex dynamic models (CAEDYM, CE-QUAL-W2, Delft 3D-ECO, LakeMab, LakeWeb, MyLake, PCLake, PROTECH, SALMO), structurally dynamic models and minimal dynamic models. We also discuss a group of approaches that could all be classified as individual based: super-individual models (Piscator, Charisma), physiologically structured models, stage-structured models and traitbased models. We briefly mention genetic algorithms, neural networks, Kalman filters and fuzzy logic. Thereafter, we zoom in, as an in-depth example, on the multi-decadal development and application of the lake ecosystem model PCLake and related models (PCLake Metamodel, Lake Shira Model, IPH-TRIM3D-PCLake). In the discussion, we argue that while the historical development of each approach and model is understandable given its 'leading principle', there are many opportunities for combining approaches. We take the point of view that a single 'right' approach does not exist and should not be strived for. Instead, multiple modelling approaches, applied concurrently to a given problem, can help develop an integrative view on the functioning of lake ecosystems. We end with a set of specific recommendations that may be of help in the further development of lake ecosystem models.
Inhalt: Introduction: The problem at hand Approaches to EU’s external identity making Mechanisms of external identity making Theoretical approaches to the EU’s external identity making The EU’s external identity promotion The ENP policy instruments Conclusions References
Die besondere Beziehung zwischen Humboldt und Darwin, zwei der bedeutendsten Persönlichkeiten in der Welt der Naturwissenschaften und der Biologie des 19. Jahrhunderts, wird detailliert auf den verschiedenen Ebenen ihres Kontaktes analysiert, sowohl was das real stattgefundene persönliche Treffen betrifft, als auch hinsichtlich ihrer Korrespondenz und der Koinzidenz von Ideen. Dieser wechselseitige Blick zeigt uns wie sich die beiden Gelehrten gegenseitig wahrnahmen, ob sie wirklich versuchten, mit dem Paradigma ihrer bedeutenden Vorgänger zu brechen, oder ob sie lediglich schrittweise das bereits erlangte Wissen erweiterten, bis es durch die Erstellung einer genialen Idee zu einem Bruch des bisherigen Wissens kommt. Bekannt ist die wiederholte Referenz von Darwin auf die Werke Humboldts, insbesondere auf die Tagebücher des deutschen Naturwissenschaftler und seine Art der Beschreibung der amerikanischen Natur in ihrer ganzen Reichhaltigkeit. Weniger bekannt hingegen sind andere Verweise in seiner Autobiografie, sowie die wissenschaftliche Verwendung des Humboldtschen Werkes oder die Zitate in seiner Korrespondenz, die in diesem Beitrag aufgezeigt werden. Darüber hinaus wird die Verwendung der frühen Schriften von Darwin durch Humboldt in einigen seiner Publikationen, vor allem im Kosmos, erwähnt.
The aim of these lectures is a reformulation and generalization of the fundamental investigations of Alexander Bach [2, 3] on the concept of probability in the work of Boltzmann [6] in the language of modern point process theory. The dominating point of view here is its subordination under the disintegration theory of Krickeberg [14]. This enables us to make Bach's consideration much more transparent. Moreover the point process formulation turns out to be the natural framework for the applications to quantum mechanical models.
Lake ecosystems across the globe have responded to climate warming of recent decades. However, correctly attributing observed changes to altered climatic conditions is complicated by multiple anthropogenic influences on lakes. This thesis contributes to a better understanding of climate impacts on freshwater phytoplankton, which forms the basis of the food chain and decisively influences water quality. The analyses were, for the most part, based on a long-term data set of physical, chemical and biological variables of a shallow, polymictic lake in north-eastern Germany (Müggelsee), which was subject to a simultaneous change in climate and trophic state during the past three decades. Data analysis included constructing a dynamic simulation model, implementing a genetic algorithm to parameterize models, and applying statistical techniques of classification tree and time-series analysis. Model results indicated that climatic factors and trophic state interactively determine the timing of the phytoplankton spring bloom (phenology) in shallow lakes. Under equally mild spring conditions, the phytoplankton spring bloom collapsed earlier under high than under low nutrient availability, due to a switch from a bottom-up driven to a top-down driven collapse. A novel approach to model phenology proved useful to assess the timings of population peaks in an artificially forced zooplankton-phytoplankton system. Mimicking climate warming by lengthening the growing period advanced algal blooms and consequently also peaks in zooplankton abundance. Investigating the reasons for the contrasting development of cyanobacteria during two recent summer heat wave events revealed that anomalously hot weather did not always, as often hypothesized, promote cyanobacteria in the nutrient-rich lake studied. The seasonal timing and duration of heat waves determined whether critical thresholds of thermal stratification, decisive for cyanobacterial bloom formation, were crossed. In addition, the temporal patterns of heat wave events influenced the summer abundance of some zooplankton species, which as predators may serve as a buffer by suppressing phytoplankton bloom formation. This thesis adds to the growing body of evidence that lake ecosystems have strongly responded to climatic changes of recent decades. It reaches beyond many previous studies of climate impacts on lakes by focusing on underlying mechanisms and explicitly considering multiple environmental changes. Key findings show that climate impacts are more severe in nutrient-rich than in nutrient-poor lakes. Hence, to develop lake management plans for the future, limnologists need to seek a comprehensive, mechanistic understanding of overlapping effects of the multi-faceted human footprint on aquatic ecosystems.
In the high mountains of Asia, glaciers cover an area of approximately 115,000 km² and constitute one of the largest continental ice accumulations outside Greenland and Antarctica. Their sensitivity to climate change makes them valuable palaeoclimate archives, but also vulnerable to current and predicted Global Warming. This is a pressing problem as snow and glacial melt waters are important sources for agriculture and power supply of densely populated regions in south, east, and central Asia. Successful prediction of the glacial response to climate change in Asia and mitigation of the socioeconomic impacts requires profound knowledge of the climatic controls and the dynamics of Asian glaciers. However, due to their remoteness and difficult accessibility, ground-based studies are rare, as well as temporally and spatially limited. We therefore lack basic information on the vast majority of these glaciers. In this thesis, I employ different methods to assess the dynamics of Asian glaciers on multiple time scales. First, I tested a method for precise satellite-based measurement of glacier-surface velocities and conducted a comprehensive and regional survey of glacial flow and terminus dynamics of Asian glaciers between 2000 and 2008. This novel and unprecedented dataset provides unique insights into the contrasting topographic and climatic controls of glacial flow velocities across the Asian highlands. The data document disparate recent glacial behavior between the Karakoram and the Himalaya, which I attribute to the competing influence of the mid-latitude westerlies during winter and the Indian monsoon during summer. Second, I tested whether such climate-related longitudinal differences in glacial behavior also prevail on longer time scales, and potentially account for observed regionally asynchronous glacial advances. I used cosmogenic nuclide surface exposure dating of erratic boulders on moraines to obtain a glacial chronology for the upper Tons Valley, situated in the headwaters of the Ganges River. This area is located in the transition zone from monsoonal to westerly moisture supply and therefore ideal to examine the influence of these two atmospheric circulation regimes on glacial advances. The new glacial chronology documents multiple glacial oscillations during the last glacial termination and during the Holocene, suggesting largely synchronous glacial changes in the western Himalayan region that are related to gradual glacial-interglacial temperature oscillations with superimposed monsoonal precipitation changes of higher frequency. In a third step, I combine results from short-term satellite-based climate records and surface velocity-derived ice-flux estimates, with topographic analyses to deduce the erosional impact of glaciations on long-term landscape evolution in the Himalayan-Tibetan realm. The results provide evidence for the long-term effects of pronounced east-west differences in glaciation and glacial erosion, depending on climatic and topographic factors. Contrary to common belief the data suggest that monsoonal climate in the central Himalaya weakens glacial erosion at high elevations, helping to maintain a steep southern orographic barrier that protects the Tibetan Plateau from lateral destruction. The results of this thesis highlight how climatic and topographic gradients across the high mountains of Asia affect glacier dynamics on time scales ranging from 10^0 to 10^6 years. Glacial response times to climate changes are tightly linked to properties such as debris cover and surface slope, which are controlled by the topographic setting, and which need to be taken into account when reconstructing mountainous palaeoclimate from glacial histories or assessing the future evolution of Asian glaciers. Conversely, the regional topographic differences of glacial landscapes in Asia are partly controlled by climatic gradients and the long-term influence of glaciers on the topographic evolution of the orogenic system.
In the present work, we study wave phenomena in strongly nonlinear lattices. Such lattices are characterized by the absence of classical linear waves. We demonstrate that compactons – strongly localized solitary waves with tails decaying faster than exponential – exist and that they play a major role in the dynamics of the system under consideration. We investigate compactons in different physical setups. One part deals with lattices of dispersively coupled limit cycle oscillators which find various applications in natural sciences such as Josephson junction arrays or coupled Ginzburg-Landau equations. Another part deals with Hamiltonian lattices. Here, a prominent example in which compactons can be found is the granular chain. In the third part, we study systems which are related to the discrete nonlinear Schrödinger equation describing, for example, coupled optical wave-guides or the dynamics of Bose-Einstein condensates in optical lattices. Our investigations are based on a numerical method to solve the traveling wave equation. This results in a quasi-exact solution (up to numerical errors) which is the compacton. Another ansatz which is employed throughout this work is the quasi-continuous approximation where the lattice is described by a continuous medium. Here, compactons are found analytically, but they are defined on a truly compact support. Remarkably, both ways give similar qualitative and quantitative results. Additionally, we study the dynamical properties of compactons by means of numerical simulation of the lattice equations. Especially, we concentrate on their emergence from physically realizable initial conditions as well as on their stability due to collisions. We show that the collisions are not exactly elastic but that a small part of the energy remains at the location of the collision. In finite lattices, this remaining part will then trigger a multiple scattering process resulting in a chaotic state.
Nanofibrous mats are interesting scaffold materials for biomedical applications like tissue engineering due to their interconnectivity and their size dimension which mimics the native cell environment. Electrospinning provides a simple route to access such fiber meshes. This thesis addresses the structural and functional control of electrospun fiber mats. In the first section, it is shown that fiber meshes with bimodal size distribution could be obtained in a single-step process by electrospinning. A standard single syringe set-up was used to spin concentrated poly(ε-caprolactone) (PCL) and poly(lactic-co-glycolic acid) (PLGA) solutions in chloroform and meshes with bimodal-sized fiber distribution could be directly obtained by reducing the spinning rate at elevated humidity. Scanning electron microscopy (SEM) and mercury porosity of the meshes suggested a suitable pore size distribution for effective cell infiltration. The bimodal fiber meshes together with unimodal fiber meshes were evaluated for cellular infiltration. While the micrometer fibers in the mixed meshes generate an open pore structure, the submicrometer fibers support cell adhesion and facilitate cell bridging on the large pores. This was revealed by initial cell penetration studies, showing superior ingrowth of epithelial cells into the bimodal meshes compared to a mesh composed of unimodal 1.5 μm fibers. The bimodal fiber meshes together with electrospun nano- and microfiber meshes were further used for the inorganic/organic hybrid fabrication of PCL with calcium carbonate or calcium phosphate, two biorelevant minerals. Such composite structures are attractive for the potential improvement of properties such as stiffness or bioactivity. It was possible to encapsulate nano and mixed sized plasma-treated PCL meshes to areas > 1 mm2 with calcium carbonate using three different mineralization methods including the use of poly(acrylic acid). The additive seemed to be useful in stabilizing amorphous calcium carbonate to effectively fill the space between the electrospun fibers resulting in composite structures. Micro-, nano- and mixed sized fiber meshes were successfully coated within hours by fiber directed crystallization of calcium phosphate using a ten-times concentrated simulated body fluid. It was shown that nanofibers accelerated the calcium phosphate crystallization, as compared to microfibers. In addition, crystallizations performed at static conditions led to hydroxyapatite formations whereas in dynamic conditions brushite coexisted. In the second section, nanofiber functionalization strategies are investigated. First, a one-step process was introduced where a peptide-polymer-conjugate (PLLA-b-CGGRGDS) was co-spun with PLGA in such a way that the peptide is enriched on the surface. It was shown that by adding methanol to the chloroform/blend solution, a dramatic increase of the peptide concentration at the fiber surface could be achieved as determined by X-ray photoelectron spectroscopy (XPS). Peptide accessibility was demonstrated via a contact angle comparison of pure PLGA and RGD-functionalized fiber meshes. In addition, the electrostatic attraction between a RGD-functionalized fiber and a silica bead at pH ~ 4 confirmed the accessibility of the peptide. The bioactivity of these RGD-functionalized fiber meshes was demonstrated using blends containing 18 wt% bioconjugate. These meshes promoted adhesion behavior of fibroblast compared to pure PLGA meshes. In a second functionalization approach, a modular strategy was investigated. In a single step, reactive fiber meshes were fabricated and then functionalized with bioactive molecules. While the electrospinning of the pure reactive polymer poly(pentafluorophenyl methacrylate) (PPFPMA) was feasible, the inherent brittleness of PPFPMA required to spin a PCL blend. Blends and pure PPFPMA showed a two-step functionalization kinetics. An initial fast reaction of the pentafluorophenyl esters with aminoethanol as a model substance was followed by a slow conversion upon further hydrophilization. This was analysed by UV/Vis-spectroscopy of the pentaflurorophenol release upon nucleophilic substitution with the amines. The conversion was confirmed by increased hydrophilicity of the resulting meshes. The PCL/PPFPMA fiber meshes were then used for functionalization with more complex molecules such as saccharides. Aminofunctionalized D-Mannose or D-Galactose was reacted with the active pentafluorophenyl esters as followed by UV/Vis spectroscopy and XPS. The functionality was shown to be bioactive using macrophage cell culture. The meshes functionalized with D-Mannose specifically stimulated the cytokine production of macrophages when lipopolysaccharides were added. This was in contrast to D-Galactose- or aminoethanol-functionalized and unfunctionalized PCL/PPFPMA fiber mats.
This thesis is concerned with the issue of extinction of populations composed of different types of individuals, and their behavior before extinction and in case of a very late extinction. We approach this question firstly from a strictly probabilistic viewpoint, and secondly from the standpoint of risk analysis related to the extinction of a particular model of population dynamics. In this context we propose several statistical tools. The population size is modeled by a branching process, which is either a continuous-time multitype Bienaymé-Galton-Watson process (BGWc), or its continuous-state counterpart, the multitype Feller diffusion process. We are interested in different kinds of conditioning on non-extinction, and in the associated equilibrium states. These ways of conditioning have been widely studied in the monotype case. However the literature on multitype processes is much less extensive, and there is no systematic work establishing connections between the results for BGWc processes and those for Feller diffusion processes. In the first part of this thesis, we investigate the behavior of the population before its extinction by conditioning the associated branching process X_t on non-extinction (X_t≠0), or more generally on non-extinction in a near future 0≤θ<∞ (X_{t+θ}≠0), and by letting t tend to infinity. We prove the result, new in the multitype framework and for θ>0, that this limit exists and is non-degenerate. This reflects a stationary behavior for the dynamics of the population conditioned on non-extinction, and provides a generalization of the so-called Yaglom limit, corresponding to the case θ=0. In a second step we study the behavior of the population in case of a very late extinction, obtained as the limit when θ tends to infinity of the process conditioned by X_{t+θ}≠0. The resulting conditioned process is a known object in the monotype case (sometimes referred to as Q-process), and has also been studied when X_t is a multitype Feller diffusion process. We investigate the not yet considered case where X_t is a multitype BGWc process and prove the existence of the associated Q-process. In addition, we examine its properties, including the asymptotic ones, and propose several interpretations of the process. Finally, we are interested in interchanging the limits in t and θ, as well as in the not yet studied commutativity of these limits with respect to the high-density-type relationship between BGWc processes and Feller processes. We prove an original and exhaustive list of all possible exchanges of limit (long-time limit in t, increasing delay of extinction θ, diffusion limit). The second part of this work is devoted to the risk analysis related both to the extinction of a population and to its very late extinction. We consider a branching population model (arising notably in the epidemiological context) for which a parameter related to the first moments of the offspring distribution is unknown. We build several estimators adapted to different stages of evolution of the population (phase growth, decay phase, and decay phase when extinction is expected very late), and prove moreover their asymptotic properties (consistency, normality). In particular, we build a least squares estimator adapted to the Q-process, allowing a prediction of the population development in the case of a very late extinction. This would correspond to the best or to the worst-case scenario, depending on whether the population is threatened or invasive. These tools enable us to study the extinction phase of the Bovine Spongiform Encephalopathy epidemic in Great Britain, for which we estimate the infection parameter corresponding to a possible source of horizontal infection persisting after the removal in 1988 of the major route of infection (meat and bone meal). This allows us to predict the evolution of the spread of the disease, including the year of extinction, the number of future cases and the number of infected animals. In particular, we produce a very fine analysis of the evolution of the epidemic in the unlikely event of a very late extinction.
This thesis is concerned with the issue of extinction of populations composed of different types of individuals, and their behavior before extinction and in case of a very late extinction. We approach this question firstly from a strictly probabilistic viewpoint, and secondly from the standpoint of risk analysis related to the extinction of a particular model of population dynamics. In this context we propose several statistical tools. The population size is modeled by a branching process, which is either a continuous-time multitype Bienaymé-Galton-Watson process (BGWc), or its continuous-state counterpart, the multitype Feller diffsion process. We are interested in different kinds of conditioning on nonextinction, and in the associated equilibrium states. These ways of conditioning have been widely studied in the monotype case. However the literature on multitype processes is much less extensive, and there is no systematic work establishing connections between the results for BGWc processes and those for Feller diffusion processes. In the first part of this thesis, we investigate the behavior of the population before its extinction by conditioning the associated branching process Xt on non-extinction (Xt 6= 0), or more generally on non-extinction in a near future 0 < 1 (Xt+ 0 = 0), and by letting t tend to infinity. We prove the result, new in the multitype framework and for 0 > 0, that this limit exists and is nondegenerate. This re ects a stationary behavior for the dynamics of the population conditioned on non-extinction, and provides a generalization of the so-called Yaglom limit, corresponding to the case 0 = 0. In a second step we study the behavior of the population in case of a very late extinction, obtained as the limit when 0 tends to infinity of the process conditioned by Xt+ 0 = 0. The resulting conditioned process is a known object in the monotype case (sometimes referred to as Q-process), and has also been studied when Xt is a multitype Feller diffusion process. We investigate the not yet considered case where Xt is a multitype BGWc process and prove the existence of the associated Q-process. In addition, we examine its properties, including the asymptotic ones, and propose several interpretations of the process. Finally, we are interested in interchanging the limits in t and 0, as well as in the not yet studied commutativity of these limits with respect to the high-density-type relationship between BGWc processes and Feller processes. We prove an original and exhaustive list of all possible exchanges of limit (long-time limit in t, increasing delay of extinction 0, diffusion limit). The second part of this work is devoted to the risk analysis related both to the extinction of a population and to its very late extinction. We consider a branching population model (arising notably in the epidemiological context) for which a parameter related to the first moments of the offspring distribution is unknown. We build several estimators adapted to different stages of evolution of the population (phase growth, decay phase, and decay phase when extinction is expected very late), and prove moreover their asymptotic properties (consistency, normality). In particular, we build a least squares estimator adapted to the Q-process, allowing a prediction of the population development in the case of a very late extinction. This would correspond to the best or to the worst-case scenario, depending on whether the population is threatened or invasive. These tools enable us to study the extinction phase of the Bovine Spongiform Encephalopathy epidemic in Great Britain, for which we estimate the infection parameter corresponding to a possible source of horizontal infection persisting after the removal in 1988 of the major route of infection (meat and bone meal). This allows us to predict the evolution of the spread of the disease, including the year of extinction, the number of future cases and the number of infected animals. In particular, we produce a very fine analysis of the evolution of the epidemic in the unlikely event of a very late extinction.
Abstract interpretation-based model checking provides an approach to verifying properties of infinite-state systems. In practice, most previous work on abstract model checking is either restricted to verifying universal properties, or develops special techniques for temporal logics such as modal transition systems or other dual transition systems. By contrast we apply completely standard techniques for constructing abstract interpretations to the abstraction of a CTL semantic function, without restricting the kind of properties that can be verified. Furthermore we show that this leads directly to implementation of abstract model checking algorithms for abstract domains based on constraints, making use of an SMT solver.
Coupling of the electrical, mechanical and optical response in polymer/liquid-crystal composites
(2010)
Micrometer-sized liquid-crystal (LC) droplets embedded in a polymer matrix may enable optical switching in the composite film through the alignment of the LC director along an external electric field. When a ferroelectric material is used as host polymer, the electric field generated by the piezoelectric effect can orient the director of the LC under an applied mechanical stress, making these materials interesting candidates for piezo-optical devices. In this work, polymer-dispersed liquid crystals (PDLCs) are prepared from poly(vinylidene fluoride-trifluoroethylene) (P(VDF-TrFE)) and a nematic liquid crystal (LC). The anchoring effect is studied by means of dielectric relaxation spectroscopy. Two dispersion regions are observed in the dielectric spectra of the pure P(VDF-TrFE) film. They are related to the glass transition and to a charge-carrier relaxation, respectively. In PDLC films containing 10 and 60 wt% LC, an additional, bias-field-dependent relaxation peak is found that can be attributed to the motion of LC molecules. Due to the anchoring effect of the LC molecules, this relaxation process is slowed down considerably, when compared with the related process in the pure LC. The electro-optical and piezo-optical behavior of PDLC films containing 10 and 60 wt% LCs is investigated. In addition to the refractive-index mismatch between the polymer matrix and the LC molecules, the interaction between the polymer dipoles and the LC molecules at the droplet interface influences the light-scattering behavior of the PDLC films. For the first time, it was shown that the electric field generated by the application of a mechanical stress may lead to changes in the transmittance of a PDLC film. Such a piezo-optical PDLC material may be useful e.g. in sensing and visualization applications. Compared to a non-polar matrix polymer, the polar matrix polymer exhibits a strong interaction with the LC molecules at the polymer/LC interface which affects the electro-optical effect of the PDLC films and prevents a larger increase in optical transmission.
Crustal deformation can be the result of volcanic and tectonic activity such as fault dislocation and magma intrusion. The crustal deformation may precede and/or succeed the earthquake occurrence and eruption. Mitigating the associated hazard, continuous monitoring of the crustal deformation accordingly has become an important task for geo-observatories and fast response systems. Due to highly non-linear behavior of the crustal deformation fields in time and space, which are not always measurable using conventional geodetic methods (e.g., Leveling), innovative techniques of monitoring and analysis are required. In this thesis I describe novel methods to improve the ability for precise and accurate mapping the spatiotemporal surface deformation field using multi acquisitions of satellite radar data. Furthermore, to better understand the source of such spatiotemporal deformation fields, I present novel static and time dependent model inversion approaches. Almost any interferograms include areas where the signal decorrelates and is distorted by atmospheric delay. In this thesis I detail new analysis methods to reduce the limitations of conventional InSAR, by combining the benefits of advanced InSAR methods such as the permanent scatterer InSAR (PSI) and the small baseline subsets (SBAS) with a wavelet based data filtering scheme. This novel InSAR time series methodology is applied, for instance, to monitor the non-linear deformation processes at Hawaii Island. The radar phase change at Hawaii is found to be due to intrusions, eruptions, earthquakes and flank movement processes and superimposed by significant environmental artifacts (e.g., atmospheric). The deformation field, I obtained using the new InSAR analysis method, is in good agreement with continuous GPS data. This provides an accurate spatiotemporal deformation field at Hawaii, which allows time dependent source modeling. Conventional source modeling methods usually deal with static deformation field, while retrieving the dynamics of the source requires more sophisticated time dependent optimization approaches. This problem I address by combining Monte Carlo based optimization approaches with a Kalman Filter, which provides the model parameters of the deformation source consistent in time. I found there are numerous deformation sources at Hawaii Island which are spatiotemporally interacting, such as volcano inflation is associated to changes in the rifting behavior, and temporally linked to silent earthquakes. I applied these new methods to other tectonic and volcanic terrains, most of which revealing the importance of associated or coupled deformation sources. The findings are 1) the relation between deep and shallow hydrothermal and magmatic sources underneath the Campi Flegrei volcano, 2) gravity-driven deformation at Damavand volcano, 3) fault interaction associated with the 2010 Haiti earthquake, 4) independent block wise flank motion at the Hilina Fault system, Kilauea, and 5) interaction between salt diapir and the 2005 Qeshm earthquake in southern Iran. This thesis, written in cumulative form including 9 manuscripts published or under review in peer reviewed journals, improves the techniques for InSAR time series analysis and source modeling and shows the mutual dependence between adjacent deformation sources. These findings allow more realistic estimation of the hazard associated with complex volcanic and tectonic systems.
rezensiertes Werk: Grossman, David: Eine Frau flieht vor einer Nachricht. - München : Hanser, 2009. - 728 S. ISBN 978-3-446-23397-3
Recent large earthquakes put in evidence the need of improving and developing robust and rapid procedures to properly calculate the magnitude of an earthquake in a short time after its occurrence. The most famous example is the 26 December 2004 Sumatra earthquake, when the limitations of the standard procedures adopted at that time by many agencies failed to provide accurate magnitude estimates of this exceptional event in time to launch early enough warnings and appropriate response. Being related to the radiated seismic energy ES, the energy magnitude ME is a good estimator of the high frequency content radiated by the source which goes into the seismic waves. However, a procedure to rapidly determine ME (that is to say, within 15 minutes after the earthquake occurrence) was required. Here it is presented a procedure able to provide in a rapid way the energy magnitude ME for shallow earthquakes by analyzing teleseismic P‑waves in the distance range 20-98. To account for the energy loss experienced by the seismic waves from the source to the receivers, spectral amplitude decay functions obtained from numerical simulations of Greens functions based on the average global model AK135Q are used. The proposed method has been tested using a large global dataset (~1000 earthquakes) and the obtained rapid ME estimations have been compared to other magnitude scales from different agencies. Special emphasis is given to the comparison with the moment magnitude MW, since the latter is very popular and extensively used in common seismological practice. However, it is shown that MW alone provide only limited information about the seismic source properties, and that disaster management organizations would benefit from a combined use of MW and ME in the prompt evaluation of an earthquake’s tsunami and shaking potential. In addition, since the proposed approach for ME is intended to work without knowledge of the fault plane geometry (often available only hours after an earthquake occurrence), the suitability of this method is discussed by grouping the analyzed earthquakes according to their type of mechanism (strike-slip, normal faulting, thrust faulting, etc.). No clear trend is found from the rapid ME estimates with the different fault plane solution groups. This is not the case for the ME routinely determined by the U.S. Geological Survey, which uses specific radiation pattern corrections. Further studies are needed to verify the effect of such corrections on ME estimates. Finally, exploiting the redundancy of the information provided by the analyzed dataset, the components of variance on the single station ME estimates are investigated. The largest component of variance is due to the intra-station (record-to-record) error, although the inter-station (station-to-station) error is not negligible and is of several magnitude units for some stations. Moreover, it is shown that the intra-station component of error is not random but depends on the travel path from a source area to a given station. Consequently, empirical corrections may be used to account for the heterogeneities of the real Earth not considered in the theoretical calculations of the spectral amplitude decay functions used to correct the recorded data for the propagation effects.
Development of techniques for earthquake microzonation studies in different urban environment
(2010)
The proliferation of megacities in many developing countries, and their location in areas where they are exposed to a high risk from large earthquakes, coupled with a lack of preparation, demonstrates the requirement for improved capabilities in hazard assessment, as well as the rapid adjustment and development of land-use planning. In particular, within the context of seismic hazard assessment, the evaluation of local site effects and their influence on the spatial distribution of ground shaking generated by an earthquake plays an important role. It follows that the carrying out of earthquake microzonation studies, which aim at identify areas within the urban environment that are expected to respond in a similar way to a seismic event, are essential to the reliable risk assessment of large urban areas. Considering the rate at which many large towns in developing countries that are prone to large earthquakes are growing, their seismic microzonation has become mandatory. Such activities are challenging and techniques suitable for identifying site effects within such contexts are needed. In this dissertation, I develop techniques for investigating large-scale urban environments that aim at being non-invasive, cost-effective and quickly deployable. These peculiarities allow one to investigate large areas over a relative short time frame, with a spatial sampling resolution sufficient to provide reliable microzonation. Although there is a negative trade-off between the completeness of available information and extent of the investigated area, I attempt to mitigate this limitation by combining two, what I term layers, of information: in the first layer, the site effects at a few calibration points are well constrained by analyzing earthquake data or using other geophysical information (e.g., shear-wave velocity profiles); in the second layer, the site effects over a larger areal coverage are estimated by means of single-station noise measurements. The microzonation is performed in terms of problem-dependent quantities, by considering a proxy suitable to link information from the first layer to the second one. In order to define the microzonation approach proposed in this work, different methods for estimating site effects have been combined and tested in Potenza (Italy), where a considerable amount of data was available. In particular, the horizontal-to-vertical spectral ratio computed for seismic noise recorded at different sites has been used as a proxy to combine the two levels of information together and to create a microzonation map in terms of spectral intensity ratio (SIR). In the next step, I applied this two-layer approach to Istanbul (Turkey) and Bishkek (Kyrgyzstan). A similar hybrid approach, i.e., combining earthquake and noise data, has been used for the microzonation of these two different urban environments. For both cities, after having calibrated the fundamental frequencies of resonance estimated from seismic noise with those obtained by analysing earthquakes (first layer), a fundamental frequency map has been computed using the noise measurements carried out within the town (second layer). By applying this new approach, maps of the fundamental frequency of resonance for Istanbul and Bishkek have been published for the first time. In parallel, a microzonation map in terms of SIR has been incorporated into a risk scenario for the Potenza test site by means of a dedicated regression between spectral intensity (SI) and macroseismic intensity (EMS). The scenario study confirms the importance of site effects within the risk chain. In fact, their introduction into the scenario led to an increase of about 50% in estimates of the number of buildings that would be partially or totally collapsed. Last, but not least, considering that the approach developed and applied in this work is based on measurements of seismic noise, their reliability has been assessed. A theoretical model describing the self-noise curves of different instruments usually adopted in microzonation studies (e.g., those used in Potenza, Istanbul and Bishkek) have been considered and compared with empirical data recorded in Cologne (Germany) and Gubbio (Italy). The results show that, depending on the geological and environmental conditions, the instrumental noise could severely bias the results obtained by recording and analysing ambient noise. Therefore, in this work I also provide some guidelines for measuring seismic noise.
The widespread usage of products containing volatile organic compounds (VOC) has lead to a general human exposure to these chemicals in work places or homes being suspected to contribute to the growing incidence of environmental diseases. Since the causal molecular mechanisms for the development of these disorders are not completely understood, the overall objective of this thesis was to investigate VOC-mediated molecular effects on human lung cells in vitro at VOC concentrations comparable to exposure scenarios below current occupational limits. Although differential expression of single proteins in response to VOCs has been reported, effects on complex protein networks (proteome) have not been investigated. However, this information is indispensable when trying to ascertain a mechanism for VOC action on the cellular level and establishing preventive strategies. For this study, the alveolar epithelial cell line A549 has been used. This cell line, cultured in a two-phase (air/liquid) model allows the most direct exposure and had been successfully applied for the analysis of inflammatory effects in response to VOCs. Mass spectrometric identification of 266 protein spots provided the first proteomic map of A549 cell line to this extent that may foster future work with this frequently used cellular model. The distribution of three typical air contaminants, monochlorobenzene (CB), styrene and 1,2 dichlorobenzene (1,2-DCB), between gas and liquid phase of the exposure model has been analyzed by gas chromatography. The obtained VOC partitioning was in agreement with available literature data. Subsequently the adapted in vitro system has been successfully employed to characterize the effects of the aromatic compound styrene on the proteome of A549 cells (Chapter 4). Initially, the cell toxicity has been assessed in order to ensure that most of the concentrations used in the following proteomic approach were not cytotoxic. Significant changes in abundance and phosphorylation in the total soluble protein fraction of A549 cells have been detected following styrene exposure. All proteins have been identified using mass spectrometry and the main cellular functions have been assigned. Validation experiments on protein and transcript level confirmed the results of the 2-DE experiments. From the results, two main cellular pathways have been identified that were induced by styrene: the cellular oxidative stress response combined with moderate pro-apoptotic signaling. Measurement of cellular reactive oxygen species (ROS) as well as the styrene-mediated induction of oxidative stress marker proteins confirmed the hypothesis of oxidative stress as the main molecular response mechanism. Finally, adducts of cellular proteins with the reactive styrene metabolite styrene 7,8 oxide (SO) have been identified. Especially the SO-adducts observed at both the reactive centers of thioredoxin reductase 1, which is a key element in the control of the cellular redox state, may be involved in styrene-induced ROS formation and apoptosis. A similar proteomic approach has been carried out with the halobenzenes CB and 1,2-DCB (Chapter 5). In accordance with previous findings, cell toxicity assessment showed enhanced toxicity compared to the one caused by styrene. Significant changes in abundance and phosphorylation of total soluble proteins of A549 cells have been detected following exposure to subtoxic concentrations of CB and 1,2-DCB. All proteins have been identified using mass spectrometry and the main cellular functions have been assigned. As for the styrene experiment, the results indicated two main pathways to be affected in the presence of chlorinated benzenes, cell death signaling and oxidative stress response. The strong induction of pro-apoptotic signaling has been confirmed for both treatments by detection of the cleavage of caspase 3. Likewise, the induction of redox-sensitive protein species could be correlated to an increased cellular level of ROS observed following CB treatment. Finally, common mechanisms in the cellular response to aromatic VOCs have been investigated (Chapter 6). A similar number (4.6-6.9%) of all quantified protein spots showed differential expression (p<0.05) following cell exposure to styrene, CB or 1,2-DCB. However, not more than three protein spots showed significant regulation in the same direction for all three volatile compounds: voltage-dependent anion-selective channel protein 2, peroxiredoxin 1 and elongation factor 2. However, all of these proteins are important molecular targets in stress- and cell death-related signaling pathways.
The difference-list technique is described in literature as effective method for extending lists to the right without using calls of append/3. There exist some proposals for automatic transformation of list programs into differencelist programs. However, we are interested in construction of difference-list programs by the programmer, avoiding the need of a transformation step. In [GG09] it was demonstrated, how left-recursive procedures with a dangling call of append/3 can be transformed into right-recursion using the unfolding technique. For simplification of writing difference-list programs using a new cons/2 procedure was introduced. In the present paper, we investigate how efficieny is influenced using cons/2. We measure the efficiency of procedures using accumulator technique, cons/2, DCG’s, and difference lists and compute the resulting speedup in respect to the simple procedure definition using append/3. Four Prolog systems were investigated and we found different behaviour concerning the speedup by difference lists. A result of our investigations is, that an often advice given in the literature for avoiding calls append/3 could not be confirmed in this strong formulation.
Data obtained from foreign data sources often come with only superficial structural information, such as relation names and attribute names. Other types of metadata that are important for effective integration and meaningful querying of such data sets are missing. In particular, relationships among attributes, such as foreign keys, are crucial metadata for understanding the structure of an unknown database. The discovery of such relationships is difficult, because in principle for each pair of attributes in the database each pair of data values must be compared. A precondition for a foreign key is an inclusion dependency (IND) between the key and the foreign key attributes. We present with Spider an algorithm that efficiently finds all INDs in a given relational database. It leverages the sorting facilities of DBMS but performs the actual comparisons outside of the database to save computation. Spider analyzes very large databases up to an order of magnitude faster than previous approaches. We also evaluate in detail the effectiveness of several heuristics to reduce the number of necessary comparisons. Furthermore, we generalize Spider to find composite INDs covering multiple attributes, and partial INDs, which are true INDs for all but a certain number of values. This last type is particularly relevant when integrating dirty data as is often the case in the life sciences domain - our driving motivation.
This professorial dissertation thesis collects several empirical studies on tax distribution and tax reform in Germany. Chapter 2 deals with two studies on effective income taxation, based on representative micro data sets from tax statistics. The first study analyses the effective income taxation at the individual level, in particular with respect to the top incomes. It is based on an integrated micro data file of household survey data and income tax statistics, which captures the entire income distribution up to the very top. Despite substantial tax base erosion and reductions of top tax rates, the German personal income tax has remained effectively progressive. The distribution of the tax burden is highly concentrated and the German economic elite is still taxed relatively heavily, even though the effective tax rate for this group has significantly declined. The second study of Chapter 2 highlights the effective income taxation of functional income sources, such as labor income, business and capital income, etc. Using income tax micro data and microsimulation models, we allocate the individual income tax liability to the respective income sources, according to different apportionment schemes accounting for losses. We find that the choice of the apportionment scheme markedly affects the tax shares of income sources and implicit tax rates, in particular those of capital income. Income types without significant losses such as labor income or transfer incomes show higher tax shares and implicit tax rates if we account for losses. The opposite is true for capital income, in particular for income from renting and leasing. Chapter 3 presents two studies on business taxation, based on representative micro data sets from tax statistics and the microsimulation model BizTax. The first part provides a study on fundamental reform options for the German local business tax. We find that today’s high concentration of local business tax revenues on corporations with high profits decreases if the tax base is broadened by integrating more taxpayers and by including more elements of business value added. The reform scenarios with a broader tax base distribute the local business tax revenue per capita more equally across regional categories. The second study of Chapter 3 discusses the macroeconomic performance of business taxation against the background of corporate income. A comparison of the tax base reported in tax statistics with the macroeconomic corporate income from national accounts gives hints to considerable tax base erosion. The average implicit tax rate on corporate income was around 20 percent since 2001, and thus falling considerably short of statutory tax rates and effective tax rates discussed in the literature. For lack of detailed accounting data it is hard to give precise reasons for the presumptive tax base erosion. Chapter 4 deals with several assessment studies on the ecological tax reform implemented in Germany as of 1999. First, we describe the scientific, ideological, and political background of the ecological tax reform. Further, we present the main findings of a first systematic impact analysis. We employ two macroeconomic models, an econometric input-output model and a recursive-dynamic computable general equilibrium (CGE) model. Both models show that Germany’s ecological tax reform helps to reduce energy consumption and CO2 emissions without having a substantial adverse effect on overall economic growth. It could have a slightly positive effect on employment. The reform’s impact on the business sector and the effects of special provisions granted to agriculture and the goods and materials sectors are outlined in a further study. The special provisions avoid higher tax burdens on the energy-intensive production. However, they widely reduce the marginal tax rates and thus the incentives to energy saving. Though the reform of special provisions 2003 increased the overall tax burden of the energy-intensive industry, the enlarged eligibility for tax rebates neutralizes the ecologic incentives. Based on the Income and Consumption Survey of 2003, we have analyzed the distributional impact of the ecological tax reform. The increased energy taxes show a clear regressive impact relative to disposable income. Families with children face a higher tax burden relative to household income. The reduction of pension contributions and the automatic adjustment of social security transfers widely mitigate this regressive impact. Households with low income or with many children nevertheless bear a slight increase in tax burden. Refunding the eco tax revenue by an eco bonus would make the reform clearly progressive.
The aim of this thesis is the design, expression and purification of human cytochrome c mutants and their characterization with regard to electrochemical and structural properties as well as with respect to the reaction with the superoxide radical and the selected proteins sulfite oxidase from human and fungi bilirubin oxidase. All three interaction partners are studied here for the first time with human cyt c and with mutant forms of cyt c. A further aim is the incorporation of the different cyt c forms in two bioelectronic systems: an electrochemical superoxide biosensor with an enhanced sensitivity and a protein multilayer assembly with and without bilirubin oxidase on electrodes. The first part of the thesis is dedicated to the design, expression and characterization of the mutants. A focus is here the electrochemical characterization of the protein in solution and immobilized on electrodes. Further the reaction of these mutants with superoxide was investigated and the possible reaction mechanisms are discussed. In the second part of the work an amperometric superoxide biosensor with selected human cytochrome c mutants was constructed and the performance of the sensor electrodes was studied. The human wild-type and four of the five mutant electrodes could be applied successfully for the detection of the superoxide radical. In the third part of the thesis the reaction of horse heart cyt c, the human wild-type and seven human cyt c mutants with the two proteins sulfite oxidase and bilirubin oxidase was studied electrochemically and the influence of the mutations on the electron transfer reactions was discussed. Finally protein multilayer electrodes with different cyt form including the mutant forms G77K and N70K which exhibit different reaction rates towards BOD were investigated and BOD together with the wild-type and engineered cyt c was embedded in the multilayer assembly. The relevant electron transfer steps and the kinetic behavior of the multilayer electrodes are investigated since the functionality of electroactive multilayer assemblies with incorporated redox proteins is often limited by the electron transfer abilities of the proteins within the multilayer. The formation via the layer-by-layer technique and the kinetic behavior of the mono and bi-protein multilayer system are studied by SPR and cyclic voltammetry. In conclusion this thesis shows that protein engineering is a helpful instrument to study protein reactions as well as electron transfer mechanisms of complex bioelectronic systems (such as bi-protein multilayers). Furthermore, the possibility to design tailored recognition elements for the construction of biosensors with an improved performance is demonstrated.
Estimation and testing the effect of covariates in accelerated life time models under censoring
(2010)
The accelerated lifetime model is considered. To test the influence of the covariate we transform the model in a regression model. Since censoring is allowed this approach leads to a goodness-of-fit problem for regression functions under censoring. So nonparametric estimation of regression functions under censoring is investigated, a limit theorem for a L2-distance is stated and a test procedure is formulated. Finally a Monte Carlo procedure is proposed.
The aim of this paper is to build and compare estimators of the infection parameter in the different phases of an epidemic (growth and extinction phases). The epidemic is modeled by a Markovian process of order d > 1 (allowing non-Markovian life spans), and can be written as a multitype branching process. We propose three estimators suitable for the different classes of criticality of the process, in particular for the subcritical case corresponding to the extinction phase. We prove their consistency and asymptotic normality for two asymptotics, when the number of ancestors (resp. number of generations) tends to infinity. We illustrate the asymptotic properties with simulated examples, and finally use our estimators to study the infection intensity in the extinction phase of the BSE epidemic in Great-Britain.
We present an approach that provides automatic or semi-automatic support for evolution and change management in heterogeneous legacy landscapes where (1) legacy heterogeneous, possibly distributed platforms are integrated in a service oriented fashion, (2) the coordination of functionality is provided at the service level, through orchestration, (3) compliance and correctness are provided through policies and business rules, (4) evolution and correctness-by-design are supported by the eXtreme Model Driven Development paradigm (XMDD) offered by the jABC (Margaria and Steffen in Annu. Rev. Commun. 57, 2004)—the model-driven service oriented development platform we use here for integration, design, evolution, and governance. The artifacts are here semantically enriched, so that automatic synthesis plugins can field the vision of Enterprise Physics: knowledge driven business process development for the end user.
We demonstrate this vision along a concrete case study that became over the past three years a benchmark for Semantic Web Service discovery and mediation. We enhance the Mediation Scenario of the Semantic Web Service Challenge along the 2 central evolution paradigms that occur in practice: (a) Platform migration: platform substitution of a legacy system by an ERP system and (b) Backend extension: extension of the legacy Customer Relationship Management (CRM) and Order Management System (OMS) backends via an additional ERP layer.
Models employed in exercise psychology highlight the role of reflective processes for explaining behavior change. However, as discussed in social cognition literature, information-processing models also consider automatic processes (dual-process models). To examine the relevance of automatic processing in exercise psychology, we used a priming task to assess the automatic evaluations of exercise stimuli in physically active sport and exercise majors (n = 32), physically active nonsport majors (n = 31), and inactive students (n = 31). Results showed that physically active students responded faster to positive words after exercise primes, whereas inactive students responded more rapidly to negative words. Priming task reaction times were successfully used to predict reported amounts of exercise in an ordinal regression model. Findings were obtained only with experiential items reflecting negative and positive consequences of exercise. The results illustrate the potential importance of dual-process models in exercise psychology.
In this paper we consider a simple syntactic extension of Answer Set Programming (ASP) for dealing with (nested) existential quantifiers and double negation in the rule bodies, in a close way to the recent proposal RASPL-1. The semantics for this extension just resorts to Equilibrium Logic (or, equivalently, to the General Theory of Stable Models), which provides a logic-programming interpretation for any arbitrary theory in the syntax of Predicate Calculus. We present a translation of this syntactic class into standard logic programs with variables (either disjunctive or normal, depending on the input rule heads), as those allowed by current ASP solvers. The translation relies on the introduction of auxiliary predicates and the main result shows that it preserves strong equivalence modulo the original signature.
Roughly every third Wikipedia article contains an infobox - a table that displays important facts about the subject in attribute-value form. The schema of an infobox, i.e., the attributes that can be expressed for a concept, is defined by an infobox template. Often, authors do not specify all template attributes, resulting in incomplete infoboxes. With iPopulator, we introduce a system that automatically populates infoboxes of Wikipedia articles by extracting attribute values from the article's text. In contrast to prior work, iPopulator detects and exploits the structure of attribute values for independently extracting value parts. We have tested iPopulator on the entire set of infobox templates and provide a detailed analysis of its effectiveness. For instance, we achieve an average extraction precision of 91% for 1,727 distinct infobox template attributes.
The seismically active Alborz mountains of northern Iran are an integral part of the Arabia-Eurasia collision. Linked strike-slip and thrust/reverse-fault systems in this mountain belt are characterized by slow loading rates, and large earthquakes are highly disparate in space and time. Similar to other intracontinental deformation zones such a pattern of tectonic activity is still insufficiently understood, because recurrence intervals between seismic events may be on the order of thousands of years, and are thus beyond the resolution of short term measurements based on GPS or instrumentally recorded seismicity. This study bridges the gap of deformation processes on different time scales. In particular, my investigation focuses on deformation on the Quaternary time scale, beyond present-day deformation rates, and it uses present-day and paleotectonic characteristics to model fault behavior. The study includes data based on structural and geomorphic mapping, faultkinematic analysis, DEM-based morphometry, and numerical fault-interaction modeling. In order to better understand the long- to short term behavior of such complex fault systems, I used geomorphic surfaces as strain markers and dated fluvial and alluvial surfaces using terrestrial cosmogenic nuclides (TCN, 10Be, 26Al, 36Cl) and optically stimulated luminescence (OSL). My investigation focuses on the seismically active Mosha-Fasham fault (MFF) and the seismically virtually inactive North Tehran Thrust (NTT), adjacent to the Tehran metropolitan area. Fault-kinematic data reveal an early mechanical linkage of the NTT and MFF during an earlier dextral transpressional stage, when the shortening direction was oriented northwest. This regime was superseded by Pliocene to Recent NE-oriented shortening, which caused thrusting and sinistral strike-slip faulting. In the course of this kinematic changeover, the NTT and MFF were reactivated and incorporated into a nascent transpressional duplex, which has significantly affected landscape evolution in this part of the range. Two of three distinctive features which characterize topography and relief in the study area can be directly related to their location inside the duplex array and are thus linked to interaction between eastern MFF and NTT, and between western MFF and Taleghan fault, respectively. To account for inferred inherited topography from the previous dextral-transpression regime, a new concept of tectonic landscape characterization has been used. Accordingly, I define simple landscapes as those environments, which have developed during the influence of a sustained tectonic regime. In contrast, composite landscapes contain topographic elements inherited from previous tectonic conditions that are inconsistent with the regional present-day stress field and kinematic style. Using numerical fault-interaction modeling with different tectonic boundary conditions, I calculated synoptic snapshots of artificial topography to compare it with the real topographic metrics. However, in the Alborz mountains, E-W faults are favorably oriented to accommodate the entire range of NW- to NE-directed compression. These faults show the highest total displacement which might indicate sustained faulting under changing boundary conditions. In contrast to the fault system within and at the flanks of the Alborz mountains, Quaternary deformation in the adjacent Tehran plain is characterized by oblique motion and thrust and strike-slip fault systems. In this morphotectonic province fault-propagation folding along major faults, limited strike-slip motion, and en-échelon arrays of second-order upper plate thrusts are typical. While the Tehran plain is characterized by young deformation phenomena, the majority of faulting took place in the early stages of the Quaternary and during late Pliocene time. TCN-dating, which was performed for the first time on geomorphic surfaces in the Tehran plain, revealed that the oldest two phases of alluviation (units A and B) must be older than late Pleistocene. While urban development in Tehran increasingly covers and obliterates the active fault traces, the present-day kinematic style, the vestiges of formerly undeformed Quaternary landforms, and paleo earthquake indicators from the last millennia attest to the threat that these faults and their related structures pose for the megacity.
On the example of the women’s magazines in Yiddish “Yidishe Froyenvelt” (1902- 1903), “Di Froy” (Vilnius1925-1933), “Froyen-Shtim” (Warsaw 1925) and “Di Froyen-Velt” (New York 1913) this article presents: • how feminist postulates are connected with questions of Jewish identity in a religious and political context • how the model image of a modern Jewish woman is presented • what the main spheres of feminist interests presented in the magazines are (a struggle for equal rights within the Jewish community as well as other social spheres, searching for and presenting outstanding women in the Jewish and world history, descriptions of women’s professional activities, psychological analysis of a woman's nature, establishing ties and a feeling of solidarity between women’s movements of other nations) • how the traditional women's roles are presented (mother, wife, housewife) • what degree of women’s participation in the edition of these periodicals is (a list of articles' authoresses and literature works appearing on columns of the periodicals) • whether and how a feminist discourse affects a language structure of the periodicals Comparing magazines from the beginning of the 20th century and the latter part of 1920s the article answers the question what direction did Jewish feminism evolve to and what content rose or fell in importance.
This thesis is concerned with the development of numerical methods using finite difference techniques for the discretization of initial value problems (IVPs) and initial boundary value problems (IBVPs) of certain hyperbolic systems which are first order in time and second order in space. This type of system appears in some formulations of Einstein equations, such as ADM, BSSN, NOR, and the generalized harmonic formulation. For IVP, the stability method proposed in [14] is extended from second and fourth order centered schemes, to 2n-order accuracy, including also the case when some first order derivatives are approximated with off-centered finite difference operators (FDO) and dissipation is added to the right-hand sides of the equations. For the model problem of the wave equation, special attention is paid to the analysis of Courant limits and numerical speeds. Although off-centered FDOs have larger truncation errors than centered FDOs, it is shown that in certain situations, off-centering by just one point can be beneficial for the overall accuracy of the numerical scheme. The wave equation is also analyzed in respect to its initial boundary value problem. All three types of boundaries - outflow, inflow and completely inflow that can appear in this case, are investigated. Using the ghost-point method, 2n-accurate (n = 1, 4) numerical prescriptions are prescribed for each type of boundary. The inflow boundary is also approached using the SAT-SBP method. In the end of the thesis, a 1-D variant of BSSN formulation is derived and some of its IBVPs are considered. The boundary procedures, based on the ghost-point method, are intended to preserve the interior 2n-accuracy. Numerical tests show that this is the case if sufficient dissipation is added to the rhs of the equations.
Fiscal federalism has been an important topic among public finance theorists in the last four decades. There is a series of arguments that decentralization of governments enhances growth by improving allocation efficiency. However, the empirical studies have shown mixed results for industrialized and developing countries and some of them have demonstrated that there might be a threshold level of economic development below which decentralization is not effective. Developing and transition countries have developed a variety of forms of fiscal decentralization as a possible strategy to achieve effective and efficient governmental structures. A generalized principle of decentralization due to the country specific circumstances does not exist. Therefore, decentralization has taken place in different forms in various countries at different times, and even exactly the same extent of decentralization may have had different impacts under different conditions. The purpose of this study is to investigate the current state of the fiscal decentralization in Mongolia and to develop policy recommendations for the efficient and effective intergovernmental fiscal relations system for Mongolia. Within this perspective the analysis concentrates on the scope and structure of the public sector, the expenditure and revenue assignment as well as on the design of the intergovernmental transfer and sub-national borrowing. The study is based on data for twenty-one provinces and the capital city of Mongolia for the period from 2000 to 2009. As a former socialist country Mongolia has had a highly centralized governmental sector. The result of the analysis below revealed that the Mongolia has introduced a number of decentralization measures, which followed a top down approach and were slowly implemented without any integrated decentralization strategy in the last decade. As a result Mongolia became de-concentrated state with fiscal centralization. The revenue assignment is lacking a very important element, for instance significant revenue autonomy given to sub-national governments, which is vital for the efficient service delivery at the local level. According to the current assignments of the expenditure and revenue responsibilities most of the provinces are unable to provide a certain national standard of public goods supply. Hence, intergovernmental transfers from the central jurisdiction to the sub-national jurisdictions play an important role for the equalization of the vertical and horizontal imbalances in Mongolia. The critical problem associated with intergovernmental transfers is that there is not a stable, predictable and transparent system of transfer allocation. The amount of transfers to sub-national governments is determined largely by political decisions on ad hoc basis and disregards local differences in needs and fiscal capacity. Thus a fiscal equalization system based on the fiscal needs of the provinces should be implemented. The equalization transfers will at least partly offset the regional disparities in revenues and enable the sub-national governments to provide a national minimum standard of local public goods.
Foraging in space and time
(2010)
All animals are adapted to the environmental conditions of the habitat they chose to live in. It was the aim of this PhD-project, to show which behavioral strategies are expressed as mechanisms to cope with the constraints, which contribute to the natural selection pressure acting on individuals. For this purpose, small mammals were exposed to different levels and types of predation risk while actively foraging. Individuals were either exposed to different predator types (airborne or ground) or combinations of both, or to indirect predators (nest predators). Risk was assumed to be distributed homogeneously, so changing the habitat or temporal adaptations where not regarded as potential options. Results show that wild-caught voles have strategic answers to this homogeneously distributed risk, which is perceived by tactile, olfactory or acoustic cues. Thus, they do not have to know an absolut quality (e.g., in terms of food provisioning and risk levels of all possible habitats), but they can adapt their behavior to the actual circumstances. Deriving risk uniform levels from cues and adjusting activity levels to the perceived risk is an option to deal with predators of the same size or with unforeseeable attack rates. Experiments showed that as long as there are no safe places or times, it is best to reduce activity and behave as inconspicuous as possible as long as the costs of missed opportunities do not exceed the benefits of a higher survival probability. Test showed that these costs apparently grow faster for males than for females, especially in times of inactivity. This is supported by strong predatory pressure on the most active groups of rodents (young males, sexually active or dispersers) leading to extremely female-biased operative sex ratios in natural populations. Other groups of animals, those with parental duties such as nest guarding, for example, have to deal with the actual risk in their habitat as well. Strategies to indirect predation pressure were tested by using bank vole mothers, confronted with a nest predator that posed no actual threat to themselves but to their young (Sorex araneus). They reduced travelling and concentrated their effort in the presence of shrews, independent of the different nutritional provisioning of food by varying resource levels due to the different seasons. Additionally, they exhibited nest-guarding strategies by not foraging in the vicinity of the nest site in order to reduce conspicuous scent marks. The repetition of the experiment in summer and autumn showed that changing environmental constraints can have a severe impact on results of outdoor studies. In our case, changing resource levels changed the type of interaction between the two species. The experiments show that it is important to analyze decision making and optimality models on an individual level, and, when that is not possible (maybe because of the constraints of field work), groups of animals should be classified by using the least common denominator that can be identified (such as sex, age, origin or kinship). This will control for the effects of the sex or stage of life history or the individual´s reproductive and nutritional status on decision making and will narrow the wide behavioral variability associated with the complex term of optimality.
The programmable network envisioned in the 1990s within standardization and research for the Intelligent Network is currently coming into reality using IPbased Next Generation Networks (NGN) and applying Service-Oriented Architecture (SOA) principles for service creation, execution, and hosting. SOA is the foundation for both next-generation telecommunications and middleware architectures, which are rapidly converging on top of commodity transport services. Services such as triple/quadruple play, multimedia messaging, and presence are enabled by the emerging service-oriented IPMultimedia Subsystem (IMS), and allow telecommunications service providers to maintain, if not improve, their position in the marketplace. SOA becomes the de facto standard in next-generation middleware systems as the system model of choice to interconnect service consumers and providers within and between enterprises. We leverage previous research activities in overlay networking technologies along with recent advances in network abstraction, service exposure, and service creation to develop a paradigm for a service environment providing converged Internet and Telecommunications services that we call Service Broker. Such a Service Broker provides mechanisms to combine and mediate between different service paradigms from the two domains Internet/WWW and telecommunications. Furthermore, it enables the composition of services across these domains and is capable of defining and applying temporal constraints during creation and execution time. By adding network-awareness into the service fabric, such a Service Broker may also act as a next generation network-to-service element allowing the composition of crossdomain and cross-layer network and service resources. The contribution of this research is threefold: first, we analyze and classify principles and technologies from Information Technologies (IT) and telecommunications to identify and discuss issues allowing cross-domain composition in a converging service layer. Second, we discuss service composition methods allowing the creation of converged services on an abstract level; in particular, we present a formalized method for model-checking of such compositions. Finally, we propose a Service Broker architecture converging Internet and Telecom services. This environment enables cross-domain feature interaction in services through formalized obligation policies acting as constraints during service discovery, creation, and execution time.
Video games structure play as performance in both the virtual and the physical space. On the one hand, the player encounters game worlds as virtual stages to act upon. On the other hand, the game world stages the player and re-frames the play space. This essay sets out to suggest some of the elements that are at work in this dualism of games as performative media. The two key elements here are the mediation of the game environment and the transformation of the player through virtual puppetry. Both cases will be argued with a focus on spatiality in performance.
This thesis presents methods, techniques and tools for developing three-dimensional representations of tactical intelligence assessments. Techniques from GIScience are combined with crime mapping methods. The range of methods applied in this study provides spatio-temporal GIS analysis as well as 3D geovisualisation and GIS programming. The work presents methods to enhance digital three-dimensional city models with application specific thematic information. This information facilitates further geovisual analysis, for instance, estimations of urban risks exposure. Specific methods and workflows are developed to facilitate the integration of spatio-temporal crime scene analysis results into 3D tactical intelligence assessments. Analysis comprises hotspot identification with kernel-density-estimation techniques (KDE), LISA-based verification of KDE hotspots as well as geospatial hotspot area characterisation and repeat victimisation analysis. To visualise the findings of such extensive geospatial analysis, three-dimensional geovirtual environments are created. Workflows are developed to integrate analysis results into these environments and to combine them with additional geospatial data. The resulting 3D visualisations allow for an efficient communication of complex findings of geospatial crime scene analysis.
The Antarctic plays an important role in the global climate system. On the one hand, the Antarctic Ice Sheet is the largest freshwater reservoir on Earth. On the other hand, a major proportion of the global bottom-water formation takes place in Antarctic shelf regions, forcing the global thermohaline circulation. The main goal of this dissertation is to provide new insights into the dynamics and stability of the EAIS during the Quaternary. Additionally, variations in the activity of bottom-water formation and their causes are investigated. The dissertation is a German contribution to the International Polar Year 2007/ 2008 and was funded by the ‘Deutsche Forschungsgesellschaft’ (DFG) within the scope of priority program 1158 ‘Antarctic research with comparative studies in Arctic ice regions’. During RV Polarstern expedition ANT-XXIII/9, glaciomarine sediments were recovered from the Prydz Bay-Kerguelen region. Prydz Bay is a key region for the study of East EAIS dynamics, as 16% of the EAIS are drained through the Lambert Glacier into the bay. Thereby, the glacier transports sediment into Prydz Bay which is then further distributed by calving icebergs or by current transport. The scientific approach of this dissertation is the reconstruction of past glaciomarine environments to infer on the response of the Lambert Glacier-Amery Ice Shelf system to climate shifts during the Quaternary. To characterize the depositional setting, sedimentological methods are used and statistical analyses are applied. Mineralogical and (bio)geochemical methods provide a means to reconstruct sediment provenances and to provide evidence on changes in the primary production in the surface water column. Age-depth models were constructed based on palaeomagnetic and palaeointensity measurements, diatom stratigraphy and radiocarbon dating. Sea-bed surface sediments in the investigation area show distinct variations in terms of their clay minerals and heavy-mineral assemblages. Considerable differences in the mineralogical composition of surface sediments are determined on the continental shelf. Clay minerals as well as heavy minerals provide useful parameters to differentiate between sediments which originated from erosion of crystalline rocks and sediments originating from Permo-Triassic deposits. Consequently, mineralogical parameters can be used to reconstruct the provenance of current-transported and ice-rafted material. The investigated sediment cores cover the time intervals of the last 1.4 Ma (continental slope) and the last 12.8 cal. ka BP (MacRobertson shelf). The sediment deposits were mainly influenced by glacial and oceanographic processes and further by biological activity (continental shelf), meltwater input and possibly gravitational transport. Sediments from the continental slope document two major deglacial events: the first deglaciation is associated with the mid-Pleistocene warming recognized around the Antarctic. In Prydz Bay, the Lambert Glacier-Amery Ice Shelf retreated far to the south and high biogenic productivity commenced or biogenic remains were better preserved due to increased sedimentation rates. Thereafter, stable glacial conditions continued until 400 - 500 ka BP. Calving of icebergs was restricted to the western part of the Lambert Glacier. The deeper bathymetry in this area allows for floating ice shelf even during times of decreased sea-level. Between 400 - 500 ka BP and the last interglacial (marine isotope stage 5) the glacier was more dynamic. During or shortly after the last interglacial the LAIS retreated again due to sea-level rise of 6 - 9 m. Both deglacial events correlate with a reduction in the thickness of ice masses in the Prince Charles Mountains. It indicates that a disintegration of the Amery Ice Shelf possibly led to increased drainage of ice masses from the Prydz Bay hinterland. A new end-member modelling algorithm was successfully applied on sediments from the MacRobertson shelf used to unmix the sand grain size fractions sorted by current activity and ice transport, respectively. Ice retreat on MacRobertson Shelf commenced 12.8 cal. ka BP and ended around 5.5 cal. ka BP. During the Holocene, strong fluctuations of the bottomwater activity were observed, probably related to variations of sea-ice formation in the Cape Darnley polynya. Increased activity of bottom-water flow was reconstructed at transitions from warm to cool conditions, whereas bottom-water activity receded during the mid- Holocene climate optimum. It can be concluded that the Lambert Glacier-Amery Ice Shelf system was relatively stable in terms of climate variations during the Quaternary. In contrast, bottom-water formation due to polynya activity was very sensitive to changes in atmospheric forcing and should gain more attention in future research.
Indonesia is one of the countries most prone to natural hazards. Complex interaction of several tectonic plates with high relative velocities leads to approximately two earthquakes with magnitude Mw>7 every year, being more than 15% of the events worldwide. Earthquakes with magnitude above 9 happen far more infrequently, but with catastrophic effects. The most severe consequences thereby arise from tsunamis triggered by these subduction-related earthquakes, as the Sumatra-Andaman event in 2004 showed. In order to enable efficient tsunami early warning, which includes the estimation of wave heights and arrival times, it is necessary to combine different types of real-time sensor data with numerical models of earthquake sources and tsunami propagation. This thesis was created as a result of the GITEWS project (German Indonesian Tsunami Early Warning System). It is based on five research papers and manuscripts. Main project-related task was the development of a database containing realistic earthquake scenarios for the Sunda Arc. This database provides initial conditions for tsunami propagation modeling used by the simulation system at the early warning center. An accurate discretization of the subduction geometry, consisting of 25x150 subfaults was constructed based on seismic data. Green’s functions, representing the deformational response to unit dip- and strike slip at the subfaults, were computed using a layered half-space approach. Different scaling relations for earthquake dimensions and slip distribution were implemented. Another project-related task was the further development of the ‘GPS-shield’ concept. It consists of a constellation of near field GPS-receivers, which are shown to be very valuable for tsunami early warning. The major part of this thesis is related to the geophysical interpretation of GPS data. Coseismic surface displacements caused by the 2004 Sumatra earthquake are inverted for slip at the fault. The effect of different Earth layer models is tested, favoring continental structure. The possibility of splay faulting is considered and shown to be a secondary order effect in respect to tsunamigenity for this event. Tsunami models based on source inversions are compared to satellite radar altimetry observations. Postseismic GPS time series are used to test a wide parameter range of uni- and biviscous rheological models of the asthenosphere. Steady-state Maxwell rheology is shown to be incompatible with near-field GPS data, unless large afterslip, amounting to more than 10% of the coseismic moment is assumed. In contrast, transient Burgers rheology is in agreement with data without the need for large aseismic afterslip. Comparison to postseismic geoid observation by the GRACE satellites reveals that even with afterslip, the model implementing Maxwell rheology results in amplitudes being too small, and thus supports a biviscous asthenosphere. A simple approach based on the assumption of quasi-static deformation propagation is introduced and proposed for inversion of coseismic near-field GPS time series. Application of this approach to observations from the 2004 Sumatra event fails to quantitatively reconstruct the rupture propagation, since a priori conditions are not fulfilled in this case. However, synthetic tests reveal the feasibility of such an approach for fast estimation of rupturing properties.
Halakha and Microhistory
(2010)
Shifra was a Jewish businesswoman in Moravia in the fifteenth-century. In 1452 due to financial fraud she was arrested in Brno. Her life was saved by some members of the local Jewish community, who renounced their financial claims against their Christian neighbours in the exchange of Shifra’s life. However, one member of the community consented to the agreement only on condition that the other members would pay his losses. The case was extensively discussed in the correspondence of contemporary rabbis, among them Israel Bruna and Israel Isserlein. Their letters about the Shifra-affair reveal some important characteristics of the rabbinic authority in the late medieval Ashkenaz.
It has been suggested that coronal mass ejections (CMEs) remove the magnetic he-licity of their coronal source region from the Sun. Such removal is often regarded to be necessary due to the hemispheric sign preference of the helicity, which inhibits a simple annihilation by reconnection between volumes of opposite chirality. Here we monitor the relative magnetic he-licity contained in the coronal volume of a simulated flux rope CME, as well as the upward flux of relative helicity through horizontal planes in the simulation box. The unstable and erupting flux rope carries away only a minor part of the initial relative helicity; the major part remains in the volume. This is a consequence of the requirement that the current through an expanding loop must decrease if the magnetic energy of the configuration is to decrease as the loop rises, to provide the kinetic energy of the CME.
This paper discusses different options for institutional arrangements providing network infrastructure on the basis of the ‘transaction cost economic’ approach using the example of highway infrastructure. Drawing on lessons learned from highway provision in three European countries (Italy, Poland and Spain), five models of highway provision are discussed: public authorities, public enterprises, user clubs, private partnerships or a regulated private market. Three options to regulate the private market are presented: a rate-of-return regulation, a price-cap-regulation and franchise bidding. The main factor that makes private construction and provision expensive are the risk premiums of private companies that are incorporated for political risks. It is argued that the optimal model of highway provision depends on each country-specific situation. This is mainly influenced by the regulatory experience within the country on one hand and by the stage of highway development on the other.
How much is too much?
(2010)
Although dietary nutrient intake is often adequate, nutritional supplement use is common among elite athletes. However, high-dose supplements or the use of multiple supplements may exceed the recommended daily allowance (RDA) of particular nutrients or even result in a daily intake above tolerable upper limits (UL). The present case report presents nutritional intake data and supplement use of a highly trained male swimmer competing at international level. Habitual energy and micronutrient intake were analysed by 3 d dietary reports. Supplement use and dosage were assessed, and total amount of nutrient supply was calculated. Micronutrient intake was evaluated based on RDA and UL as presented by the European Scientific Committee on Food, and maximum permitted levels in supplements (MPL) are given. The athlete’s diet provided adequate micronutrient content well above RDA except for vitamin D. Simultaneous use of ten different supplements was reported, resulting in excess intake above tolerable UL for folate, vitamin E and Zn. Additionally, daily supplement dosage was considerably above MPL for nine micronutrients consumed as artificial products. Risks and possible side effects of exceeding UL by the athlete are discussed. Athletes with high energy intake may be at risk of exceeding UL of particular nutrients if multiple supplements are added. Therefore, dietary counselling of athletes should include assessment of habitual diet and nutritional supplement intake. Educating athletes to balance their diets instead of taking supplements might be prudent to prevent health risks
that may occur with long-term excess nutrient intake.
Large open-source software projects involve developers with a wide variety of backgrounds and expertise. Such software projects furthermore include many internal APIs that developers must understand and use properly. According to the intended purpose of these APIs, they are more or less frequently used, and used by developers with more or less expertise. In this paper, we study the impact of usage patterns and developer expertise on the rate of defects occurring in the use of internal APIs. For this preliminary study, we focus on memory management APIs in the Linux kernel, as the use of these has been shown to be highly error prone in previous work. We study defect rates and developer expertise, to consider e.g., whether widely used APIs are more defect prone because they are used by less experienced developers, or whether defects in widely used APIs are more likely to be fixed.
In a very simplified view, the plant leaf growth can be reduced to two processes, cell division and cell expansion, accompanied by expansion of their surrounding cell walls. The vacuole, as being the largest compartment of the plant cell, plays a major role in controlling the water balance of the plant. This is achieved by regulating the osmotic pressure, through import and export of solutes over the vacuolar membrane (the tonoplast) and by controlling the water channels, the aquaporins. Together with the control of cell wall relaxation, vacuolar osmotic pressure regulation is thought to play an important role in cell expansion, directly by providing cell volume and indirectly by providing ion and pH homestasis for the cytosoplasm. In this thesis the role of tonoplast protein coding genes in cell expansion in the model plant Arabidopsis thaliana is studied and genes which play a putative role in growth are identified. Since there is, to date, no clearly identified protein localization signal for the tonoplast, there is no possibility to perform genome-wide prediction of proteins localized to this compartment. Thus, a series of recent proteomic studies of the tonoplast were used to compile a list of cross-membrane tonoplast protein coding genes (117 genes), and other growth-related genes from notably the growth regulating factor (GRF) and expansin families were included (26 genes). For these genes a platform for high-throughput reverse transcription quantitative real time polymerase chain reaction (RT-qPCR) was developed by selecting specific primer pairs. To this end, a software tool (called QuantPrime, see http://www.quantprime.de) was developed that automatically designs such primers and tests their specificity in silico against whole transcriptomes and genomes, to avoid cross-hybridizations causing unspecific amplification. The RT-qPCR platform was used in an expression study in order to identify candidate growth related genes. Here, a growth-associative spatio-temporal leaf sampling strategy was used, targeting growing regions at high expansion developmental stages and comparing them to samples taken from non-expanding regions or stages of low expansion. Candidate growth related genes were identified after applying a template-based scoring analysis on the expression data, ranking the genes according to their association with leaf expansion. To analyze the functional involvement of these genes in leaf growth on a macroscopic scale, knockout mutants of the candidate growth related genes were screened for growth phenotypes. To this end, a system for non-invasive automated leaf growth phenotyping was established, based on a commercially available image capture and analysis system. A software package was developed for detailed developmental stage annotation of the images captured with the system, and an analysis pipeline was constructed for automated data pre-processing and statistical testing, including modeling and graph generation, for various growth-related phenotypes. Using this system, 24 knockout mutant lines were analyzed, and significant growth phenotypes were found for five different genes.
Russian Jews who left the Former Soviet Union (FSU) and its Successor States after 1989 are considered as one of the best qualified migrants group worldwide. In the preferred countries of destination (Israel, the United States and Germany) they are well-known for cultural self-assertion, strong social upward mobility and manifold forms of self organisation and empowerment. Using Suzanne Kellers sociological model of “Strategic Elites”, it easily becomes clear that a huge share of the Russian Jewish Immigrants in Germany and Israel are part of various elites due to their qualification and high positions in the FSU – first of all professional, cultural and intellectual elites (“Intelligentsija”). The study aimed to find out to what extent developments of cultural self-assertion, of local and transnational networking and of ethno-cultural empowerment are supported or even initiated by the immigrated (Russian Jewish) Elites. The empirical basis for this study have been 35 half-structured expert interviews with Russian Jews in both countries (Israel, Germany) – most of them scholars, artists, writers, journalists/publicists, teachers, engineers, social workers, students and politicians. The qualitative analysis of the interview material in Israel and Germany revealed that there are a lot of commonalities but also significant differences. It was obvious that almost all of the interview partners remained to be linked with Russian speaking networks and communities, irrespective of their success (or failure) in integration into the host societies. Many of them showed self-confidence with regard to the groups’ amazing professional resources (70% of the adults with academic degree), and the cultural, professional and political potential of the FSU immigrants was usually considered as equal to those of the host population(s). Thus, the immigrants’ interest in direct societal participation and social acceptance was accordingly high. Assimilation was no option. For the Russian Jewish “sense of community” in Israel and Germany, Russian Language, Arts and general Russian culture have remained of key importance. The Immigrants do not feel an insuperable contradiction when feeling “Russian” in cultural terms, “Jewish” in ethnical terms and “Israeli” / “German” in national terms – in that a typical case of additive identity shaping what is also significant for the Elites of these Immigrants. Tendencies of ethno-cultural self organisation – which do not necessarily hinder impressing individual careers in the new surroundings – are more noticeable in Israel. Thus, a part of the Russian Jewish Elites has responded to social exclusion, discrimination or blocking by local population (and by local elites) with intense efforts to build (Russian Jewish) Associations, Media, Educational Institutions and even Political Parties. All in all, the results of this study do very much contradict popular stereotypes of the Russian Jewish Immigrant as a pragmatic, passive “Homo Sovieticus”. Among the Interview Partners in this study, civil-societal commitment was not the exception but rather the rule. Traditional activities of the early, legendary Russian „Intelligentsija“ were marked by smooth transitions from arts, education and societal/political commitment. There seem to be certain continuities of this self-demand in some of the Russian Jewish groups in Israel. Though, nothing comparable could be drawn from the Interviews with the Immigrants in Germany. Thus, the myth and self-demand of Russian “Intelligentsija” is irrelevant for collective discourses among Russian Jews in Germany.
Background
Micrometer resolution placement and immobilization of probe molecules is an important step in the preparation of biochips and a wide range of lab-on-chip systems. Most known methods for such a deposition of several different substances are costly and only suitable for a limited number of probes. In this article we present a flexible procedure for simultaneous spatially controlled immobilization of functional biomolecules by molecular ink lithography.
Results
For the bottom-up fabrication of surface bound nanostructures a universal method is presented that allows the immobilization of different types of biomolecules with micrometer resolution. A supporting surface is biotinylated and streptavidin molecules are deposited with an AFM (atomic force microscope) tip at distinct positions. Subsequent incubation with a biotinylated molecule species leads to binding only at these positions. After washing streptavidin is deposited a second time with the same AFM tip and then a second biotinylated molecule species is coupled by incubation. This procedure can be repeated several times. Here we show how to immobilize different types of biomolecules in an arbitrary arrangement whereas most common methods can deposit only one type of molecules. The presented method works on transparent as well as on opaque substrates. The spatial resolution is better than 400 nm and is limited only by the AFM's positional accuracy after repeated z-cycles since all steps are performed in situ without moving the supporting surface. The principle is demonstrated by hybridization to different immobilized DNA oligomers and was validated by fluorescence microscopy.
Conclusions
The immobilization of different types of biomolecules in high-density microarrays is a challenging task for biotechnology. The method presented here not only allows for the deposition of DNA at submicrometer resolution but also for proteins and other molecules of biological relevance that can be coupled to biotin.
Flood design necessitates discharge estimates for large recurrence intervals. However, in a flood frequency analysis, the uncertainty of discharge estimates increases with higher recurrence intervals, particularly due to the small number of available flood data. Furthermore, traditional distribution functions increase unlimitedly without consideration of an upper bound discharge. Hence, additional information needs to be considered which is representative for high recurrence intervals. Envelope curves which bound the maximum observed discharges of a region are an adequate regionalisation method to provide additional spatial information for the upper tail of a distribution function. Probabilistic regional envelope curves (PRECs) are an extension of the traditional empirical envelope curve approach, in which a recurrence interval is estimated for a regional envelope curve (REC). The REC is constructed for a homogeneous pooling group of sites. The estimation of this recurrence interval is based on the effective sample years of data considering the intersite dependence among all sites of the pooling group. The core idea of this thesis was an improvement of discharge estimates for high recurrence intervals by integrating empirical and probabilistic regional envelope curves into the flood frequency analysis. Therefore, the method of probabilistic regional envelope curves was investigated in detail. Several pooling groups were derived by modifying candidate sets of catchment descriptors and settings of two different pooling methods. These were used to construct PRECs. A sensitivity analysis shows the variability of discharges and the recurrence intervals for a given site due to the different assumptions. The unit flood of record which governs the intercept of PREC was determined as the most influential aspect. By separating the catchments into nested and unnested pairs, the calculation algorithm for the effective sample years of data was refined. In this way, the estimation of the recurrence intervals was improved, and therefore the use of different parameter sets for nested and unnested pairs of catchments is recommended. In the second part of this thesis, PRECs were introduced into a distribution function. Whereas in the traditional approach only discharge values are used, PRECs provide a discharge and its corresponding recurrence interval. Hence, a novel approach was developed, which allows a combination of the PREC results with the traditional systematic flood series while taking the PREC recurrence interval into consideration. An adequate mixed bounded distribution function was presented, which in addition to the PREC results also uses an upper bound discharge derived by an empirical envelope curve. By doing so, two types of additional information which are representative for the upper tail of a distribution function were included in the flood frequency analysis. The integration of both types of additional information leads to an improved discharge estimation for recurrence intervals between 100 and 1000 years.
This thesis contains quantum chemical models and force field calculations for the RuBisCO isotope effect, the spectral characteristics of the blue-light sensor BLUF and the light harvesting complex II. The work focuses on the influence of the environment on the corresponding systems. For RuBisCO, it was found that the isotopic effect is almost unaffected by the environment. In case of the BLUF domain, an amino acid was found to be important for the UV/vis spectrum, but unaccounted for in experiments so far (Ser41). The residue was shown to be highly mobile and with a systematic influence on the spectral shift of the BLUF domain chromophore (flavin). Finally, for LHCII it was found that small changes in the geometry of a Chlorophyll b/Violaxanthin chromophore pair can have strong influences regarding the light harvesting mechanism. Especially here it was seen that the proper description of the environment can be critical. In conclusion, the environment was observed to be of often unexpected importance for the molecular properties, and it seems not possible to give a reliable estimate on the changes created by the presence of the environment.
The origin and evolution of granites has been widely studied because granitoid rocks constitute a major portion of the Earth ́s crust. The formation of granitic magma is, besides temperature mainly triggered by the water content of these rocks. The presence of water in magmas plays an important role due to the ability of aqueous fluids to change the chemical composition of the magma. The exsolution of aqueous fluids from melts is closely linked to a fractionation of elements between the two phases. Then, aqueous fluids migrate to shallower parts of the Earth ́s crust because of it ́s lower density compared to that of melts and adjacent rocks. This process separates fluids and melts, and furthermore, during the ascent, aqueous fluids can react with the adjacent rocks and alter their chemical signature. This is particularly impor- tant during the formation of magmatic-hydrothermal ore deposits or in the late stages of the evolution of magmatic complexes. For a deeper insight to these processes, it is essential to improve our knowledge on element behavior in such systems. In particular, trace elements are used for these studies and petrogenetic interpretations because, unlike major elements, they are not essential for the stability of the phases involved and often reflect magmatic processes with less ambiguity. However, for the majority of important trace elements, the dependence of the geochemical behavior on temperature, pressure, and in particular on the composition of the system are only incompletely or not at all experimentally studied. Former studies often fo- cus on the determination of fluid−melt partition coefficients (Df/m=cfluid/cmelt) of economically interesting elements, e.g., Mo, Sn, Cu, and there are some partitioning data available for ele- ments that are also commonly used for petrological interpretations. At present, no systematic experimental data on trace element behavior in fluid−melt systems as function of pressure, temperature, and chemical composition are available. Additionally, almost all existing data are based on the analysis of quenched phases. This results in substantial uncertainties, particularly for the quenched aqueous fluid because trace element concentrations may change upon cooling. The objective of this PhD thesis consisted in the study of fluid−melt partition coefficients between aqueous solutions and granitic melts for different trace elements (Rb, Sr, Ba, La, Y, and Yb) as a function of temperature, pressure, salinity of the fluid, composition of the melt, and experimental and analytical approach. The latter included the refinement of an existing method to measure trace element concentrations in fluids equilibrated with silicate melts di- rectly at elevated pressures and temperatures using a hydrothermal diamond-anvil cell and synchrotron radiation X-ray fluorescence microanalysis. The application of this in-situ method enables to avoid the main source of error in data from quench experiments, i.e., trace element concentration in the fluid. A comparison of the in-situ results to data of conventional quench experiments allows a critical evaluation of quench data from this study and literature data. In detail, starting materials consisted of a suite of trace element doped haplogranitic glasses with ASI varying between 0.8 and 1.4 and H2O or a chloridic solution with m NaCl/KCl=1 and different salinities (1.16 to 3.56 m (NaCl+KCl)). Experiments were performed at 750 to 950◦C and 0.2 or 0.5 GPa using conventional quench devices (externally and internally heated pressure vessels) with different quench rates, and at 750◦C and 0.2 to 1.4 GPa with in-situ analysis of the trace element concentration in the fluids. The fluid−melt partitioning data of all studied trace elements show 1. a preference for the melt (Df/m < 1) at all studied conditions, 2. one to two orders of magnitude higher Df/m using chloridic solutions compared to experiments with H2O, 3. a clear dependence on the melt composition for fluid−melt partitioning of Sr, Ba, La, Y, and Yb in experiments using chloridic solutions, 4. quench rate−related differences of fluid−melt partition coefficients of Rb and Sr, and 5. distinctly higher fluid−melt partitioning data obtained from in-situ experiments than from comparable quench runs, particularly in the case of H2O as starting solution. The data point to a preference of all studied trace elements for the melt even at fairly high salinities, which contrasts with other experimental studies, but is supported by data from studies of natural co-genetically trapped fluid and melt inclusions. The in-situ measurements of trace element concentrations in the fluid verify that aqueous fluids will change their composition upon cooling, which is in particular important for Cl free systems. The distinct differences of the in-situ results to quench data of this study as well as to data from the literature signify the im- portance of a careful fluid sampling and analysis. Therefore, the direct measurement of trace element contents in fluids equilibrated with silicate melts at elevated PT conditions represents an important development to obtain more reliable fluid−melt partition coefficients. For further improvement, both the aqueous fluid and the silicate melt need to be analyzed in-situ because partitioning data that are based on the direct measurement of the trace element content in the fluid and analysis of a quenched melt are still not completely free of quench effects. At present, all available data on element complexation in aqueous fluids in equilibrium with silicate melts at high PT are indirectly derived from partitioning data, which involves in these experiments assumptions on the species present in the fluid. However, the activities of chemical components in these partitioning experiments are not well constrained, which is required for the definition of exchange equilibria between melt and fluid species. For example, the melt-dependent variation of partition coefficient observed for Sr imply that this element can not only be complexed by Cl− as suggested previously. The data indicate a more complicated complexation of Sr in the aqueous fluid. To verify this hypothesis, the in-situ setup was also used to determine strontium complexation in fluids equilibrated with silicate melts at desired PT conditions by the application of X-ray absorption near edge structure (XANES) spectroscopy. First results show a strong effect of both fluid and melt composition on the resulting XANES spectra, which indicates different complexation environments for Sr.
Inverse agonist and neutral antagonist actions of synthetic compounds at an insect 5-HT1 receptor
(2010)
Background and purpose: 5-Hydroxytryptamine (5-HT) has been shown to control and modulate many physiological and behavioural functions in insects. In this study, we report the cloning and pharmacological properties of a 5-HT1 receptor of an insect model for neurobiology, physiology and pharmacology. Experimental approach: A cDNA encoding for the Periplaneta americana 5-HT1 receptor was amplified from brain cDNA. The receptor was stably expressed in HEK 293 cells, and the functional and pharmacological properties were determined in cAMP assays. Receptor distribution was investigated by RT-PCR and by immunocytochemistry using an affinity-purified polyclonal antiserum. Key results: The P. americana 5-HT1 receptor (Pea5-HT1) shares pronounced sequence and functional similarity with mammalian 5-HT1 receptors. Activation with 5-HT reduced adenylyl cyclase activity in a dose-dependent manner. Pea5-HT1 was expressed as a constitutively active receptor with methiothepin acting as a neutral antagonist, and WAY 100635 as an inverse agonist. Receptor mRNA was present in various tissues including brain, salivary glands and midgut. Receptor-specific antibodies showed that the native protein was expressed in a glycosylated form in membrane samples of brain and salivary glands. Conclusions and implications: This study marks the first pharmacological identification of an inverse agonist and a neutral antagonist at an insect 5-HT1 receptor. The results presented here should facilitate further analyses of 5-HT1 receptors in mediating central and peripheral effects of 5-HT in insects.
Fire prone Mediterranean-type vegetation systems like those in the Mediterranean Basin and South-Western Australia are global hot spots for plant species diversity. To ensure management programs act to maintain these highly diverse plant communities, it is necessary to get a profound understanding of the crucial mechanisms of coexistence. In the current literature several mechanisms are discussed. The objective of my thesis is to systematically explore the importance of potential mechanisms for maintaining multi-species, fire prone vegetation by modelling. The model I developed is spatially-explicit, stochastic, rule- and individual-based. It is parameterised on data of population dynamics collected over 18 years in the Mediterranean-type shrublands of Eneabba, Western Australia. From 156 woody species of the area seven plant traits have been identified to be relevant for this study: regeneration mode, annual maximum seed production, seed size, maximum crown diameter, drought tolerance, dispersal mode and seed bank type. Trait sets are used for the definition of plant functional types (PFTs). The PFT dynamics are simulated annual by iterating life history processes. In the first part of my thesis I investigate the importance of trade-offs for the maintenance of high diversity in multi-species systems with 288 virtual PFTs. Simulation results show that the trade-off concept can be helpful to identify non-viable combinations of plant traits. However, the Shannon Diversity Index of modelled communities can be high despite of the presence of ‘supertypes’. I conclude, that trade-offs between two traits are less important to explain multi-species coexistence and high diversity than it is predicted by more conceptual models. Several studies show, that seed immigration from the regional seed pool is essential for maintaining local species diversity. However, systematical studies on the seed rain composition to multi-species communities are missing. The results of the simulation experiments, as presented in part two of this thesis, show clearly, that without seed immigration the local species community found in Eneabba drifts towards a state with few coexisting PFTs. With increasing immigration rates the number of simulated coexisting PFTs and Shannon diversity quickly approaches values as also observed in the field. Including the regional seed input in the model is suited to explain more aggregated measures of the local plant community structure such as species richness and diversity. Hence, the seed rain composition should be implemented in future studies. In the third part of my thesis I test the sensitivity of Eneabba PFTs to four different climate change scenarios, considering their impact on both local and regional processes. The results show that climate change clearly has the potential to alter the number of dispersed seeds for most of the Eneabba PFTs and therefore the source of the ‘immigrants’ at the community level. A classification tree analysis shows that, in general, the response to climate change was PFT-specific. In the Eneabba sand plains sensitivity of a PFT to climate change depends on its specific trait combination and on the scenario of environmental change i.e. development of the amount of rainfall and the fire frequency. This result emphasizes that PFT-specific responses and regional process seed immigration should not be ignored in studies dealing with the impact of climate change on future species distribution. The results of the three chapters are finally analysed in a general discussion. The model is discussed and improvements and suggestions are made for future research. My work leads to the following conclusions: i) It is necessary to support modelling with empirical work to explain coexistence in species-rich plant communities. ii) The chosen modelling approach allows considering the complexity of coexistence and improves the understanding of coexistence mechanisms. iii) Field research based assumptions in terms of environmental conditions and plant life histories can relativise the importance of more hypothetic coexistence theories in species-rich systems. In consequence, trade-offs can play a lower role than predicted by conceptual models. iv) Seed immigration is a key process for local coexistence. Its alteration because of climate change should be considered for prognosis of coexistence. Field studies should be carried out to get data on seed rain composition.
rezensiertes Werk: Gartner, Isabella: Menorah : Jüdisches Familienblatt für Wissenschaft, Kunst und Literatur (1923–1932) ; Materialien zur Geschichte einer Wiener zionistischen Zeitschrift. - Würzburg : Königshausen & Neumann, 2009. - 356 S. ISBN 978-3-8260-3864-8
We present the tool Kato which is, to the best of our knowledge, the first tool for plagiarism detection that is directly tailored for answer-set programming (ASP). Kato aims at finding similarities between (segments of) logic programs to help detecting cases of plagiarism. Currently, the tool is realised for DLV programs but it is designed to handle various logic-programming syntax versions. We review basic features and the underlying methodology of the tool.
Left peripheral focus
(2010)
In Czech, German, and many other languages, part of the semantic focus
of the utterance can be moved to the left periphery of the clause. The main generalization is that only the leftmost accented part of the semantic focus can be moved. We propose that movement to the left periphery is generally triggered by an unspecific edge feature of C (Chomsky 2008) and its restrictions can be attributed to requirements of cyclic linearization, modifying the theory of cyclic linearization developed by Fox and Pesetsky (2005). The crucial assumption is that structural accent is a direct consequence of being linearized at merge, thus it is indirectly relevant for (locality restrictions on) movement. The absence of structural accent correlates with given-ness. Given elements may later receive (topic or contrastive) accents, which accounts for fronting in multiple focus/contrastive topic constructions. Without any additional assumptions, the model can account for movement of pragmatically unmarked elements to the left periphery (‘formal fronting’, Frey 2005). Crucially, the analysis makes no reference at all to concepts of information structure in the syntax, in line with the claim of Chomsky (2008) that UG specifies no direct link between syntax and information structure.
rezensiertes Werk: Leshonot yehude Sefarad ve-ha-mizrach vesifruyotehem / Languages and literatures of Sephardic and Oriental Jews. - Jerusalem : Misgav Yerushalayim, 2009. - 484 S. [hebr.] + 434 S. [lat.], ; Ill.
This is the 13th issue of the working paper series Interdisciplinary Studies on Information Structure (ISIS) of the Sonderforschungsbereich (SFB) 632. It is the first part of a series of Linguistic Fieldnote issues which present data collected by members of different projects of the SFB during fieldwork on various languages or dialects spoken worldwide. This part of the Fieldnote Series is dedicated to data from African languages. It contains contributions by Mira Grubic (A5) on Ngizim, and Susanne Genzel & Frank Kügler (D5) on Akan. The papers allow insights into various aspects of the elicitation of formal correlates of focus and related phenomena in different African languages investigated by the SFB in the second funding phase, especially in the period between 2007 and 2010.
Live cell flattening
(2010)
Eukaryotic cell flattening is valuable for improving microscopic observations, ranging from bright field (BF) to total internal reflection fluorescence (TIRF) microscopy. Fundamental processes, such as mitosis and in vivo actin polymerization, have been investigated using these techniques. Here, we review the well known agar overlayer protocol and the oil overlay method. In addition, we present more elaborate microfluidics-based techniques that provide us with a greater level of control. We demonstrate these techniques on the social amoebae Dictyostelium discoideum, comparing the advantages and disadvantages of each method.
Background: Local adaptation to divergent environmental conditions can promote population genetic differentiation even in the absence of geographic barriers and hence, lead to speciation. Perturbations by catastrophic events, however, can distort such parapatric ecological speciation processes. Here, we asked whether an exceptionally strong flood led to homogenization of gene pools among locally adapted populations of the Atlantic molly (Poecilia mexicana, Poeciliidae) in the Cueva del Azufre system in southern Mexico, where two strong environmental selection factors (darkness within caves and/or presence of toxic H2S in sulfidic springs) drive the diversification of P. mexicana. Nine nuclear microsatellites as well as heritable female life history traits (both as a proxy for quantitative genetics and for trait divergence) were used as markers to compare genetic differentiation, genetic diversity, and especially population mixing (immigration and emigration) before and after the flood. Results: Habitat type (i.e., non-sulfidic surface, sulfidic surface, or sulfidic cave), but not geographic distance was the major predictor of genetic differentiation. Before and after the flood, each habitat type harbored a genetically distinct population. Only a weak signal of individual dislocation among ecologically divergent habitat types was uncovered (with the exception of slightly increased dislocation from the Cueva del Azufre into the sulfidic creek, El Azufre). By contrast, several lines of evidence are indicative of increased flood-induced dislocation within the same habitat type, e.g., between different cave chambers of the Cueva del Azufre. Conclusions: The virtual absence of individual dislocation among ecologically different habitat types indicates strong natural selection against migrants. Thus, our current study exemplifies that ecological speciation in this and other systems, in which extreme environmental factors drive speciation, may be little affected by temporary perturbations, as adaptations to physico-chemical stressors may directly affect the survival probability in divergent habitat types.
The fourth volume of the DIGAREC Series holds the proceedings to the conference “Logic and Structure of the Computer Game”, held at the House of Brandenburg- Prussian History in Potsdam on November 6 and 7, 2009. The conference was the first to explicitly address the medial logic and structure of the computer game. The contributions focus on the specific potential for mediation and on the unique form of mediation inherent in digital games. This includes existent, yet scattered approaches to develop a unique curriculum of game studies. In line with the concept of ‘mediality’, the notions of aesthetics, interactivity, software architecture, interface design, iconicity, spatiality, and rules are of special interest. Presentations were given by invited German scholars and were commented on by international respondents in a dialogical structure.
This paper comprises four parts. Firstly, an overview of the mathematics of decision logic in relation to games and of the construction of narration and characters is given. This includes specific limits of the use of decision logic pertaining to games in general and to storytelling in particular. Secondly, the rule system as the medial unconsciousness is focused on. Thirdly, remarks are made on the debate between ludology and narratology, which had to fail as it missed the crucial point: the computer game as a medium. Finally, gaming in general, as well as its relationship to chance, coincidence, emergence, and event is discussed.
Logic as a medium
(2010)
Computer games are rigid in a peculiar way: the logic of computation was the first to shape the early games. The logic of interactivity marked the action genre of games in the second place, while in massive multiplayer online gaming all the emergences of the net occur to confront us with just another type of logic. These logics are the media in which the specific forms of computer games evolve. Therefore, a look at gaming supposing that there are three eras of computation is taken: the early synthetical era, ruled by the Turing machine and by mainframe computers, by the IPO principle of computing; the second, mimetical era, when interactivity and graphical user interfaces dominate, the domain of the feedback loop; and the third, emergent era, in which the complexity of networked personal computers and their users is dominant.
An important characteristic of Service-Oriented Architectures is that clients do not depend on the service implementation's internal assignment of methods to objects. It is perhaps the most important technical characteristic that differentiates them from more common object-oriented solutions. This characteristic makes clients and services malleable, allowing them to be rearranged at run-time as circumstances change. That improvement in malleability is impaired by requiring clients to direct service requests to particular services. Ideally, the clients are totally oblivious to the service structure, as they are to aspect structure in aspect-oriented software. Removing knowledge of a method implementation's location, whether in object or service, requires re-defining the boundary line between programming language and middleware, making clearer specification of dependence on protocols, and bringing the transaction-like concept of failure scopes into language semantics as well. This paper explores consequences and advantages of a transition from object-request brokering to service-request brokering, including the potential to improve our ability to write more parallel software.
CHAMP (CHAllenging Minisatellite Payload) is a German small satellite mission to study the earth's gravity field, magnetic field and upper atmosphere. Thanks to the good condition of the satellite so far, the planned 5 years mission is extended to year 2009. The satellite provides continuously a large quantity of measurement data for the purpose of Earth study. The measurements of the magnetic field are undertaken by two Fluxgate Magnetometers (vector magnetometer) and one Overhauser Magnetometer (scalar magnetometer) flown on CHAMP. In order to ensure the quality of the data during the whole mission, the calibration of the magnetometers has to be performed routinely in orbit. The scalar magnetometer serves as the magnetic reference and its readings are compared with the readings of the vector magnetometer. The readings of the vector magnetometer are corrected by the parameters that are derived from this comparison, which is called the scalar calibration. In the routine processing, these calibration parameters are updated every 15 days by means of scalar calibration. There are also magnetic effects coming from the satellite which disturb the measurements. Most of them have been characterized during tests before launch. Among them are the remanent magnetization of the spacecraft and fields generated by currents. They are all considered to be constant over the mission life. The 8 years of operation experience allow us to investigate the long-term behaviors of the magnetometers and the satellite systems. According to the investigation, it was found that for example the scale factors of the FGM show obvious long-term changes which can be described by logarithmic functions. The other parameters (offsets and angles between the three components) can be considered constant. If these continuous parameters are applied for the FGM data processing, the disagreement between the OVM and the FGM readings is limited to \pm1nT over the whole mission. This demonstrates, the magnetometers on CHAMP exhibit a very good stability. However, the daily correction of the parameter Z component offset of the FGM improves the agreement between the magnetometers markedly. The Z component offset plays a very important role for the data quality. It exhibits a linear relationship with the standard deviation of the disagreement between the OVM and the FGM readings. After Z offset correction, the errors are limited to \pm0.5nT (equivalent to a standard deviation of 0.2nT). We improved the corrections of the spacecraft field which are not taken into account in the routine processing. Such disturbance field, e.g. from the power supply system of the satellite, show some systematic errors in the FGM data and are misinterpreted in 9-parameter calibration, which brings false local time related variation of the calibration parameters. These corrections are made by applying a mathematical model to the measured currents. This non-linear model is derived from an inversion technique. If the disturbance field of the satellite body are fully corrected, the standard deviation of scalar error \triangle B remains about 0.1nT. Additionally, in order to keep the OVM readings a reliable standard, the imperfect coefficients of the torquer current correction for the OVM are redetermined by solving a minimization problem. The temporal variation of the spacecraft remanent field is investigated. It was found that the average magnetic moment of the magneto-torquers reflects well the moment of the satellite. This allows for a continuous correction of the spacecraft field. The reasons for the possible unknown systemic error are discussed in this thesis. Particularly, both temperature uncertainties and time errors have influence on the FGM data. Based on the results of this thesis the data processing of future magnetic missions can be designed in an improved way. In particular, the upcoming ESA mission Swarm can take advantage of our findings and provide all the auxiliary measurements needed for a proper recovery of the ambient magnetic field.
The Mariae Vitae Congregation was the first and possibly the most important missionary institution in Poland-Lithuania in the eighteenth and nineteenth centuries. According to the Rule of Mariae Vitae Congregation, it had to deal with religious and lay education of converted girls (mainly Jewish) and provide them with practical skills of work so they could establish in Catholic society. The innovatory social program of Mariae Vitae Congregation including education and financial help answered to possible problems of neophytes in Poland and Lithuania of that time.
Two recent magnetic field models, GRIMM and xCHAOS, describe core field accelerations with similar behavior up to Spherical Harmonic (SH) degree 5, but which differ significantly for higher degrees. These discrepancies, due to different approaches in smoothing rapid time variations of the core field, have strong implications for the interpretation of the secular variation. Furthermore, the amount of smoothing applied to the highest SH degrees is essentially the modeler’s choice. We therefore investigate new ways of regularizing core magnetic field models. Here we propose to constrain field models to be consistent with the frozen flux induction equation by co-estimating a core magnetic field model and a flow model at the top of the outer core. The flow model is required to have smooth spatial and temporal behavior. The implementation of such constraints and their effects on a magnetic field model built from one year of CHAMP satellite and observatory data, are presented. In particular, it is shown that the chosen constraints are efficient and can be used to build reliable core magnetic field secular variation and acceleration model components.
The debate whether to locate the narrative of digital games a) as part of the code or b) as part of the performance will be the starting point for an analysis of two roleplaying games: the single-player game ZELDA: MAJORA’S MASK and the Korean MMORPG AION and their respective narrative logics. When we understand games as abstract code systems, then the narrative logic can be understood as embedded on the code level. With a focus on the player’s performance, the actualization of the possibilities given in the code system is central. Both logics, that of code and that of performance, are reflected in players’ narratives based on the playing experience. They do reflect on the underlying code and rules of the game system as they do reflect on the game world and their own performance within. These narratives rely heavily on the source text – the digital game –, which means that they give insights into the underlying logics of the source text. I will discuss the game structure, the players’ performance while playing the game and the performance of the player after playing the game producing fan narratives. I conceive the narrative structure and the performance of the player playing as necessarily interconnected when we discuss the narrative logics of a game. Producing fan narratives is understood as a performance as well. This performance is based on the experience the players made while playing and refers to both logics of the game they use as their source text.
The availability of large data sets has allowed researchers to uncover complex properties in complex systems, such as complex networks and human dynamics. A vast number of systems, from the Internet to the brain, power grids, ecosystems, can be represented as large complex networks. Dynamics on and of complex networks has attracted more and more researchers’ interest. In this thesis, first, I introduced a simple but effective dynamical optimization coupling scheme which can realize complete synchronization in networks with undelayed and delayed couplings and enhance the small-world and scale-free networks’ synchronizability. Second, I showed that the robustness of scale-free networks with community structure was enhanced due to the existence of communities in the networks and some of the response patterns were found to coincide with topological communities. My results provide insights into the relationship between network topology and the functional organization in complex networks from another viewpoint. Third, as an important kind of nodes of complex networks, human detailed correspondence dynamics was studied by both data and the model. A new and general type of human correspondence pattern was found and an interacting priority-queues model was introduced to explain it. The model can also embrace a range of realistic social interacting systems such as email and letter communication. My findings provide insight into various human activities both at the individual and network level. Fourth, I present clearly new evidence that human comment behavior in on-line social systems, a different type of interacting human dynamics, is non-Poissonian and a model based on the personal attraction was introduced to explain it. These results are helpful for discovering regular patterns of human behavior in on-line society and the evolution of the public opinion on the virtual as well as real society. Finally, there are conclusion and outlook of human dynamics and complex networks.
On a method for solution of the ordinary differential equations connected with Huygens' equations
(2010)
This paper is a critical examination of the relationship between reality and simulation. After a brief theoretical introduction, it unfolds its argument on an empirical level, using a thick game playing description of GRAND THEFT AUTO IV. This in-game experience serves as material for the subsequent analysis, in the course of which defining characteristics of computer game playing are formulated. Finally, on the basis of this analysis, the paper postulates the hypothesis that playing computer games like GTA IV promotes competency in deconstructing simulations and implements a cyclic logic of recreation.
A constraint programming system combines two essential components: a constraint solver and a search engine. The constraint solver reasons about satisfiability of conjunctions of constraints, and the search engine controls the search for solutions by iteratively exploring a disjunctive search tree defined by the constraint program. The Monadic Constraint Programming framework gives a monadic definition of constraint programming where the solver is defined as a monad threaded through the monadic search tree. Search and search strategies can then be defined as firstclass objects that can themselves be built or extended by composable search transformers. Search transformers give a powerful and unifying approach to viewing search in constraint programming, and the resulting constraint programming system is first class and extremely flexible.
Parafoveal Load of Word N+1 Modulates Preprocessing Effectivenessof Word N+2 in Chinese Reading
(2010)
Preview benefits (PBs) from two words to the right of the fixated one (i.e., word N+2)and associated parafoveal-on-foveal effects are critical for proposals of distributed lexical processing during reading. This experiment examined parafoveal processing during reading of Chinese sentences, using a boundary manipulation of N+2-word preview with low- and high-frequency words N+1. The main findings were (a) an identity PB for word N+2 that was (b) primarily observed when word N+1 was of high frequency (i.e., an interaction between frequency of word N+1 and PB for word N+2), and (c) a parafoveal-on-foveal frequency effect of word N+1 for fixation durations on word N. We discuss implications for theories of serial attention shifts and parallel distributed processing of words during reading.
Pattern matching is a well-established concept in the functional programming community. It provides the means for concisely identifying and destructuring values of interest. This enables a clean separation of data structures and respective functionality, as well as dispatching functionality based on more than a single value. Unfortunately, expressive pattern matching facilities are seldomly incorporated in present object-oriented programming languages. We present a seamless integration of pattern matching facilities in an object-oriented and dynamically typed programming language: Newspeak. We describe language extensions to improve the practicability and integrate our additions with the existing programming environment for Newspeak. This report is based on the first author’s master’s thesis.
In the most abstract definition of its operational semantics, the declarative and concurrent programming language CHR is trivially non-terminating for a significant class of programs. Common refinements of this definition, in closing the gap to real-world implementations, compromise on declarativity and/or concurrency. Building on recent work and the notion of persistent constraints, we introduce an operational semantics avoiding trivial non-termination without compromising on its essential features.
The Takab complex is composed of a variety of metamorphic rocks including amphibolites, metapelites, mafic granulites, migmatites and meta-ultramafics, which are intruded by the granitoid. The granitoid magmatic activity occurred in relation to the subduction of the Neo-Tethys oceanic crust beneath the Iranian crust during Tertiary times. The granitoids are mainly granodiorite, quartz monzodiorite, monzonite and quartz diorite. Chemically, the magmatic rocks are characterized by ASI < 1.04, AI < 0.87 and high contents of CaO (up to ∼ 14.5 wt %), which are consistent with the I-type magmatic series. Low FeO t /(FeO t +MgO) values (< 0.75) as well as low Nb, Y and K 2 O contents of the investigated rocks resemble the calc-alkaline series. Low SiO 2 , K 2 O/Na 2 O and Al 2 O 3 accompanied by high CaO and FeO contents indicate melting of metabasites as an appropriate source for the intrusions. Negative Ti and Nb anomalies verify a metaluminous crustal origin for the protoliths of the investigated igneous rocks. These are comparable with compositions of the associated mafic migmatites, in the Takab metamorphic complex, which originated from the partial melting of amphibolites. Therefore, crustal melting and a collision-related origin for the Takab calc-alkaline intrusions are proposed here on the basis of mineralogy and geochemical characteristics. The P–T evolution during magmatic crystallization and subsolidus cooling stages is determined by the study of mineral chemistry of the granodiorite and the quartz diorite. Magmatic crystallization pressure and temperature for the quartz-diorite and the granodiorite are estimated to be P ∼ 7.8 ± 2.5 kbar, T ∼ 760 ± 75 ◦C and P ∼ 5 ± 1 kbar, T ∼ 700 ◦C, respectively. Subsolidus conditions are consistent with temperatures of ∼ 620 ◦C and ∼ 600 ◦C, and pressures of ∼ 5 kbar and ∼ 3.5 kbar for the quartz-diorite and the granodiorite, respectively.
This work describes the realization of physically crosslinked networks based on gelatin by the introduction of functional groups enabling specific supramolecular interactions. Molecular models were developed in order to predict the material properties and permit to establish a knowledge-based approach to material design. The effect of additional supramolecular interactions with hydroxyapaptite was then studied in composite materials. The calculated properties are compared to experimental results to validate the models. The models are then further used for the study of physically crosslinked networks. Gelatin was functionalized with desaminotyrosine (DAT) and desaminotyrosyl-tyrosine (DATT) side groups, derived from the natural amino acid tyrosine. These group can potentially undergo to π-π and hydrogen bonding interactions also under physiological conditions. Molecular dynamics (MD) simulations were performed on models with 0.8 wt.-% or 25 wt.-% water content, using the second generation forcefield CFF91. The validation of the models was obtained by the comparison with specific experimental data such as, density, peptide conformational angles and X-ray scattering spectra. The models were then used to predict the supramolecular organization of the polymer chain, analyze the formation of physical netpoints and calculate the mechanical properties. An important finding of simulation was that with the increase of aromatic groups also the number of observed physical netpoints increased. The number of relatively stable physical netpoints, on average zero 0 for natural gelatin, increased to 1 and 6 for DAT and DATT functionalized gelatins respectively. A comparison with the Flory-Rehner model suggested reduced equilibrium swelling by factor 6 of the DATT-functionalized materials in water. The functionalized gelatins could be synthesized by chemoselective coupling of the free carboxylic acid groups of DAT and DATT to the free amino groups of gelatin. At 25 wt.-% water content, the simulated and experimentally determined elastic mechanical properties (e.g. Young Modulus) were both in the order of GPa and were not influenced by the degree of aromatic modification. The experimental equilibrium degree of swelling in water decreased with increasing the number of inserted aromatic functions (from 2800 vol.-% for pure gelatin to 300 vol.-% for the DATT modified gelatin), at the same time, Young’s modulus, elongation at break, and maximum tensile strength increased. It could be show that the functionalization with DAT and DATT influences the chain organization of gelatin based materials together with a controlled drying condition. Functionalization with DAT and DATT lead to a drastic reduction of helical renaturation, that could be more finely controlled by the applied drying conditions. The properties of the materials could then be influenced by application of two independent methods. Composite materials of DAT and DATT functionalized gelatins with hydroxyapatite (HAp) show a drastic reduction of swelling degree. In tensile tests and rheological measurements, the composites equilibrated in water had increased Young’s moduli (from 200 kPa up to 2 MPa) and tensile strength (from 57 kPa up to 1.1 MPa) compared to the natural polymer matrix without affecting the elongation at break. Furthermore, an increased thermal stability from 40 °C to 85 °C of the networks could be demonstrated. The differences of the behaviour of the functionalized gelatins to pure gelatin as matrix suggested an additional stabilizing bond between the incorporated aromatic groups to the hydroxyapatite.