Filtern
Volltext vorhanden
- ja (382) (entfernen)
Erscheinungsjahr
- 2017 (382) (entfernen)
Dokumenttyp
- Postprint (237)
- Dissertation (100)
- Wissenschaftlicher Artikel (16)
- Monographie/Sammelband (8)
- Masterarbeit (4)
- Preprint (4)
- Arbeitspapier (4)
- Konferenzveröffentlichung (3)
- Ausgabe (Heft) zu einer Zeitschrift (2)
- Bericht (2)
Sprache
- Englisch (382) (entfernen)
Schlagworte
- climate (9)
- Germany (6)
- climate change (6)
- networks (5)
- variability (5)
- ancient DNA (4)
- elderly (4)
- model (4)
- precipitation (4)
- transport (4)
Institut
- Mathematisch-Naturwissenschaftliche Fakultät (72)
- Institut für Geowissenschaften (42)
- Institut für Biochemie und Biologie (39)
- Humanwissenschaftliche Fakultät (29)
- Institut für Physik und Astronomie (28)
- Institut für Chemie (24)
- Strukturbereich Kognitionswissenschaften (19)
- Department Musik und Kunst (12)
- Institut für Ernährungswissenschaft (12)
- Sozialwissenschaften (12)
Ancient genomes have revolutionized our understanding of Holocene prehistory and, particularly, the Neolithic transition in western Eurasia. In contrast, East Asia has so far received little attention, despite representing a core region at which the Neolithic transition took place independently similar to 3 millennia after its onset in the Near East. We report genome-wide data from two hunter-gatherers from Devil's Gate, an early Neolithic cave site (dated to similar to 7.7 thousand years ago) located in East Asia, on the border between Russia and Korea. Both of these individuals are genetically most similar to geographically close modern populations from the Amur Basin, all speaking Tungusic languages, and, in particular, to the Ulchi. The similarity to nearby modern populations and the low levels of additional genetic material in the Ulchi imply a high level of genetic continuity in this region during the Holocene, a pattern that markedly contrasts with that reported for Europe.
The performance of hybridization capture combined with next-generation sequencing (NGS) has seen limited investigation with samples from hot and arid regions until now. We applied hybridization capture and shotgun sequencing to recover DNA sequences from bone specimens of ancient-domestic dromedary (Camelus dromedarius) and its extinct ancestor, the wild dromedary from Jordan, Syria, Turkey and the Arabian Peninsula, respectively. Our results show that hybridization capture increased the percentage of mitochondrial DNA (mtDNA) recovery by an average 187-fold and in some cases yielded virtually complete mitochondrial (mt) genomes at multifold coverage in a single capture experiment. Furthermore, we tested the effect of hybridization temperature and time by using a touchdown approach on a limited number of samples. We observed no significant difference in the number of unique dromedary mtDNA reads retrieved with the standard capture compared to the touchdown method. In total, we obtained 14 partial mitochondrial genomes from ancient-domestic dromedaries with 17-95% length coverage and 1.27-47.1-fold read depths for the covered regions. Using whole-genome shotgun sequencing, we successfully recovered endogenous dromedary nuclear DNA (nuDNA) from domestic and wild dromedary specimens with 1-1.06-fold read depths for covered regions. Our results highlight that despite recent methodological advances, obtaining ancient DNA (aDNA) from specimens recovered from hot, arid environments is still problematic. Hybridization protocols require specific optimization, and samples at the limit of DNA preservation need multiple replications of DNA extraction and hybridization capture as has been shown previously for Middle Pleistocene specimens.
We introduce an abstract concept of quantum field theory on categories fibered in groupoids over the category of spacetimes. This provides us with a general and flexible framework to study quantum field theories defined on spacetimes with extra geometric structures such as bundles, connections and spin structures. Using right Kan extensions, we can assign to any such theory an ordinary quantum field theory defined on the category of spacetimes and we shall clarify under which conditions it satisfies the axioms of locally covariant quantum field theory. The same constructions can be performed in a homotopy theoretic framework by using homotopy right Kan extensions, which allows us to obtain first toy-models of homotopical quantum field theories resembling some aspects of gauge theories.
Background
Cells are able to communicate and coordinate their function within tissues via secreted factors. Aberrant secretion by cancer cells can modulate this intercellular communication, in particular in highly organised tissues such as the liver. Hepatocytes, the major cell type of the liver, secrete Dickkopf (Dkk), which inhibits Wnt/ β-catenin signalling in an autocrine and paracrine manner. Consequently, Dkk modulates the expression of Wnt/ β-catenin target genes. We present a mathematical model that describes the autocrine and paracrine regulation of hepatic gene expression by Dkk under wild-type conditions as well as in the presence of mutant cells.
Results
Our spatial model describes the competition of Dkk and Wnt at receptor level, intra-cellular Wnt/ β-catenin signalling, and the regulation of target gene expression for 21 individual hepatocytes. Autocrine and paracrine regulation is mediated through a feedback mechanism via Dkk and Dkk diffusion along the porto-central axis. Along this axis an APC concentration gradient is modelled as experimentally detected in liver. Simulations of mutant cells demonstrate that already a single mutant cell increases overall Dkk concentration. The influence of the mutant cell on gene expression of surrounding wild-type hepatocytes is limited in magnitude and restricted to hepatocytes in close proximity. To explore the underlying molecular mechanisms, we perform a comprehensive analysis of the model parameters such as diffusion coefficient, mutation strength and feedback strength.
Conclusions
Our simulations show that Dkk concentration is elevated in the presence of a mutant cell. However, the impact of these elevated Dkk levels on wild-type hepatocytes is confined in space and magnitude. The combination of inter- and intracellular processes, such as Dkk feedback, diffusion and Wnt/ β-catenin signal transduction, allow wild-type hepatocytes to largely maintain their gene expression.
Background
The members of the genus Muntiacus are of particular interest to evolutionary biologists due to their extreme chromosomal rearrangements and the ongoing discussions about the number of living species. Red muntjacs have the largest distribution of all muntjacs and were formerly considered as one species. Karyotype differences led to the provisional split between the Southern Red Muntjac (Muntiacus muntjak) and the Northern Red Muntjac (M. vaginalis), but uncertainties remain as, so far, no phylogenetic study has been conducted. Here, we analysed whole mitochondrial genomes of 59 archival and 16 contemporaneous samples to resolve uncertainties about their taxonomy and used red muntjacs as model for understanding the evolutionary history of other species in Southeast Asia.
Results
We found three distinct matrilineal groups of red muntjacs: Sri Lankan red muntjacs (including the Western Ghats) diverged first from other muntjacs about 1.5 Mya; later northern red muntjacs (including North India and Indochina) and southern red muntjacs (Sundaland) split around 1.12 Mya. The diversification of red muntjacs into these three main lineages was likely promoted by two Pleistocene barriers: one through the Indian subcontinent and one separating the Indochinese and Sundaic red muntjacs. Interestingly, we found a high level of gene flow within the populations of northern and southern red muntjacs, indicating gene flow between populations in Indochina and dispersal of red muntjacs over the exposed Sunda Shelf during the Last Glacial Maximum.
Conclusions
Our results provide new insights into the evolution of species in South and Southeast Asia as we found clear genetic differentiation in a widespread and generalist species, corresponding to two known biogeographical barriers: The Isthmus of Kra and the central Indian dry zone. In addition, our molecular data support either the delineation of three monotypic species or three subspecies, but more importantly these data highlight the conservation importance of the Sri Lankan/South Indian red muntjac.
Purpose: Comparison of the dissociation kinetics of rapid-acting insulins lispro, aspart, glulisine and human insulin under physiologically relevant conditions.
Methods: Dissociation kinetics after dilution were monitored directly in terms of the average molecular mass using combined static and dynamic light scattering. Changes in tertiary structure were detected by near-UV circular dichroism.
Results: Glulisine forms compact hexamers in formulation even in the absence of Zn2+. Upon severe dilution, these rapidly dissociate into monomers in less than 10 s. In contrast, in formulations of lispro and aspart, the presence of Zn2+ and phenolic compounds is essential for formation of compact R6 hexamers. These slowly dissociate in times ranging from seconds to one hour depending on the concentration of phenolic additives. The disadvantage of the long dissociation times of lispro and aspart can be diminished by a rapid depletion of the concentration of phenolic additives independent of the insulin dilution. This is especially important in conditions similar to those after subcutaneous injection, where only minor dilution of the insulins occurs.
Conclusion: Knowledge of the diverging dissociation mechanisms of lispro and aspart compared to glulisine will be helpful for optimizing formulation conditions of rapid-acting insulins.
West German anticommunism and the SED’s Westarbeit were to some extentinterrelated. From the beginning, each German state had attemted to stabilise itsown social system while trying to discredit its political opponent. The claim tosole representation and the refusal to acknowledge each other delineated governmentalaction on both sides. Anticommunism inWest Germany re-developed under theconditions of the Cold War, which allowed it to become virtually the reason ofstate and to serve as a tool for the exclusion of KPD supporters. In its turn, theSED branded the West German State as‘revanchist’and instrumentalised itsanticommunism to persecute and eliminate opponents within the GDR. Bothphenomena had an integrative and exclusionary element.
With its transparent orthography, Standard Indonesian is spoken by over 160 million inhabitants and is the primary language of instruction in education and the government in Indonesia. An assessment battery of reading and reading-related skills was developed as a starting point for the diagnosis of dyslexia in beginner learners. Founded on the International Dyslexia Association’s definition of dyslexia, the test battery comprises nine empirically motivated reading and reading-related tasks assessing word reading, pseudoword reading, arithmetic, rapid automatized naming, phoneme deletion, forward and backward digit span, verbal fluency, orthographic choice (spelling), and writing. The test was validated by computing the relationships between the outcomes on the reading-skills and reading-related measures by means of correlation and factor analyses. External variables, i.e., school grades and teacher ratings of the reading and learning abilities of individual students, were also utilized to provide evidence of its construct validity. Four variables were found to be significantly related with reading-skill measures: phonological awareness, rapid naming, spelling, and digit span. The current study on reading development in Standard Indonesian confirms findings from other languages with transparent orthographies and suggests a test battery including preliminary norm scores for screening and assessment of elementary school children learning to read Standard Indonesian.
Many studies demonstrated interactions between number processing and either spatial codes (effects of spatial-numerical associations) or visual size-related codes (size-congruity effect). However, the interrelatedness of these two number couplings is still unclear. The present study examines the simultaneous occurrence of space- and size-numerical congruency effects and their interactions both within and across trials, in a magnitude judgment task physically small or large digits were presented left or right from screen center. The reaction times analysis revealed that space- and size-congruency effects coexisted in parallel and combined additively. Moreover, a selective sequential modulation of the two congruency effects was found. The size-congruency effect was reduced after size incongruent trials. The space-congruency effect, however, was only affected by the previous space congruency. The observed independence of spatial-numerical and within magnitude associations is interpreted as evidence that the two couplings reflect Different attributes of numerical meaning possibly related to orginality and cardinality.
Integral Fourier operators
(2017)
This volume of contributions based on lectures delivered at a school on Fourier Integral Operators
held in Ouagadougou, Burkina Faso, 14–26 September 2015, provides an introduction to Fourier Integral Operators (FIO) for a readership of Master and PhD students as well as any interested layperson. Considering the wide
spectrum of their applications and the richness of the mathematical tools they involve, FIOs lie the cross-road of many a field. This volume offers
the necessary background, whether analytic or geometric, to get acquainted with FIOs, complemented by more advanced material presenting various aspects of active research in that area.
Via their powerful radiation, stellar winds, and supernova explosions, massive stars (Mini & 8 M☉) bear a tremendous impact on galactic evolution. It became clear in recent decades that the majority of massive stars reside in binary systems. This thesis sets as a goal to quantify the impact of binarity (i.e., the presence of a companion star) on massive stars. For this purpose, massive binary systems in the Local Group, including OB-type binaries, high mass X-ray binaries (HMXBs), and Wolf-Rayet (WR) binaries, were investigated by means of spectral, orbital, and evolutionary analyses.
The spectral analyses were performed with the non-local thermodynamic equillibrium (non-LTE) Potsdam Wolf-Rayet (PoWR) model atmosphere code. Thanks to critical updates in the calculation of the hydrostatic layers, the code became a state-of-the-art tool applicable for all types of hot massive stars (Chapter 2). The eclipsing OB-type triple system δ Ori served as an intriguing test-case for the new version of the PoWR code, and provided key insights regarding the formation of X-rays in massive stars (Chapter 3). We further analyzed two prototypical HMXBs, Vela X-1 and IGR J17544-2619, and obtained fundamental conclusions regarding the dichotomy of two basic classes of HMXBs (Chapter 4). We performed an exhaustive analysis of the binary R 145 in the Large Magellanic Cloud (LMC), which was claimed to host the most massive stars known. We were able to disentangle the spectrum of the system, and performed an orbital, polarimetric, and spectral analysis, as well as an analysis of the wind-wind collision region. The true masses of the binary components turned out to be significantly lower than suggested, impacting our understanding of the initial mass function and stellar evolution at low metallicity (Chapter 5). Finally, all known WR binaries in the Small Magellanic Cloud (SMC) were analyzed. Although it was theoretical predicted that virtually all WR stars in the SMC should be formed via mass-transfer in binaries, we find that binarity was not important for the formation of the known WR stars in the SMC, implying a strong discrepancy between theory and observations (Chapter 6).
This is a brief survey of a constructive technique of analytic continuation related to an explicit integral formula of Golusin and Krylov (1933). It goes far beyond complex analysis and applies to the Cauchy problem for elliptic partial differential equations as well. As started in the classical papers, the technique is elaborated in generalised Hardy spaces also called Hardy-Smirnov spaces.
Statistics Canada, Canada’s national statistics agency, offers a suite of spatial
files for mapping and analysis of its various population data products. The following
article showcases possibilities and shortfalls of the existing spatial files
for mapping population data, and provides an overview of the structure of the
available boundary files from the regional to the dissemination block level. Due
to Canada’s highly dispersed population, mapping its distribution and density can
be challenging. Common mapping techniques such as the choropleth method are
suitable only for mapping spatially high resolution data such as data at the dissemination
area level. To allow for mapping of population data at less detailed levels
such as census divisions or provinces, Statistics Canada has created a so-called
ecumene boundary file which outlines the inhabited area of Canada and can be
used to more accurately visualize Canada’s population distribution and density.
Earth observation data have become an outstanding basis for analyzing environmental
aspects. The increasing availability of remote sensing data is accompanied
by an increasing user demand. Within the scope of the COOPERNICUS-initiative,
the automatic processing of remote sensing data is important for supplying value-
added-information products. The use of additional data like land-water-masks
in the context of deriving value-added information products can stabilize and
improve the product quality of information products.
The authors of this contribution would like to discuss different automated
processing algorithms which are based on land-water masks for value-added
data interpretation. These developments were supported or accompanied by Prof.
Hartmut Asche.
Mental arithmetic exhibits various biases. Among those is a tendency to overestimate addition and to underestimate subtraction outcomes. Does such “operational momentum” (OM) also affect multiplication and division? Twenty-six adults produced lines whose lengths corresponded to the correct outcomes of multiplication and division problems shown in symbolic format. We found a reliable tendency to over-estimate division outcomes, i.e., reverse OM. We suggest that anchoring on the first operand (a tendency to use this number as a reference for further quantitative reasoning) contributes to cognitive biases in mental arithmetic.
Trunk loading and back pain
(2017)
An essential function of the trunk is the compensation of external forces and loads in order to guarantee stability. Stabilising the trunk during sudden, repetitive loading in everyday tasks, as well as during performance is important in order to protect against injury. Hence, reduced trunk stability is accepted as a risk factor for the development of back pain (BP). An altered activity pattern including extended response and activation times as well as increased co-contraction of the trunk muscles as well as a reduced range of motion and increased movement variability of the trunk are evident in back pain patients (BPP). These differences to healthy controls (H) have been evaluated primarily in quasi-static test situations involving isolated loading directly to the trunk. Nevertheless, transferability to everyday, dynamic situations is under debate. Therefore, the aim of this project is to analyse 3-dimensional motion and neuromuscular reflex activity of the trunk as response to dynamic trunk loading in healthy (H) and back pain patients (BPP).
A measurement tool was developed to assess trunk stability, consisting of dynamic test situations. During these tests, loading of the trunk is generated by the upper and lower limbs with and without additional perturbation. Therefore, lifting of objects and stumbling while walking are adequate represents. With the help of a 12-lead EMG, neuromuscular activity of the muscles encompassing the trunk was assessed. In addition, three-dimensional trunk motion was analysed using a newly developed multi-segmental trunk model. The set-up was checked for reproducibility as well as validity. Afterwards, the defined measurement set-up was applied to assess trunk stability in comparisons of healthy and back pain patients.
Clinically acceptable to excellent reliability could be shown for the methods (EMG/kinematics) used in the test situations. No changes in trunk motion pattern could be observed in healthy adults during continuous loading (lifting of objects) of different weights. In contrast, sudden loading of the trunk through perturbations to the lower limbs during walking led to an increased neuromuscular activity and ROM of the trunk. Moreover, BPP showed a delayed muscle response time and extended duration until maximum neuromuscular activity in response to sudden walking perturbations compared to healthy controls. In addition, a reduced lateral flexion of the trunk during perturbation could be shown in BPP.
It is concluded that perturbed gait seems suitable to provoke higher demands on trunk stability in adults. The altered neuromuscular and kinematic compensation pattern in back pain patients (BPP) can be interpreted as increased spine loading and reduced trunk stability in patients. Therefore, this novel assessment of trunk stability is suitable to identify deficits in BPP. Assignment of affected BPP to therapy interventions with focus on stabilisation of the trunk aiming to improve neuromuscular control in dynamic situations is implied. Hence, sensorimotor training (SMT) to enhance trunk stability and compensation of unexpected sudden loading should be preferred.
This thesis investigates the processing of non-canonical word orders and whether non-canonical orders involving object topicalizations, midfield scrambling and particle verbs are treated the same by native (L1) and non-native (L2) speakers. The two languages investigated are Norwegian and German.
32 L1 Norwegian and 32 L1 German advanced learners of Norwegian were tested in two experiments on object topicalization in Norwegian. The results from the online self-paced reading task and the offline agent identification task show that both groups are able to identify the non-canonical word order and show a facilitatory effect of animate subjects in their reanalysis. Similarly high error rates in the agent identification task suggest that globally unambiguous object topicalizations are a challenging structure for L1 and L2 speakers alike.
The same participants were also tested in two experiments on particle placement in Norwegian, again using a self-paced reading task, this time combined with an acceptability rating task. In the acceptability rating L1 and L2 speakers show the same preference for the verb-adjacent placement of the particle over the non-adjacent placement after the direct object. However, this preference for adjacency is only found in the L1 group during online processing, whereas the L2 group shows no preference for either order.
Another set of experiments tested 33 L1 German and 39 L1 Slavic advanced learners of German on object scrambling in ditransitive clauses in German. Non-native speakers accept both object orders and show neither a preference for either order nor a processing advantage for the canonical order. The L1 group, in contrast, shows a small, but significant preference for the canonical dative-first order in the judgment and the reading task.
The same participants were also tested in two experiments on the application of the split rule in German particle verbs. Advanced L2 speakers of German are able to identify particle verbs and can apply the split rule in V2 contexts in an acceptability judgment task in the same way as L1 speakers. However, unlike the L1 group, the L2 group is not sensitive to the grammaticality manipulation during online processing. They seem to be sensitive to the additional lexical information provided by the particle, but are unable to relate the split particle to the preceding verb and recognize the ungrammaticality in non-V2 contexts.
Taken together, my findings suggest that non-canonical word orders are not per se more difficult to identify for L2 speakers than L1 speakers and can trigger the same reanalysis processes as in L1 speakers. I argue that L2 speakers’ ability to identify a non-canonical word order depends on how the non-canonicity is signaled (case marking vs. surface word order), on the constituents involved (identical vs. different word types), and on the impact of the word order change on sentence meaning. Non-canonical word orders that are signaled by morphological case marking and cause no change to the sentence’s content are hard to detect for L2 speakers.
Background
Non-typhoid Salmonella Typhimurium (S. Typhimurium) accounts for a high number of registered salmonellosis cases, and O-serotyping is one important tool for monitoring epidemiology and spread of the disease. Moreover, variations in glucosylated O-antigens are related to immunogenicity and spread in the host. However, classical autoagglutination tests combined with the analysis of specific genetic markers cannot always reliably register phase variable glucose modifications expressed on Salmonella O-antigens and additional tools to monitor O-antigen glucosylation phenotypes of S. Typhimurium would be desirable.
Results
We developed a test for the phase variable O-antigen glucosylation state of S. Typhimurium using the tailspike proteins (TSP) of Salmonella phages 9NA and P22. We used this ELISA like tailspike adsorption (ELITA) assay to analyze a library of 44 Salmonella strains. ELITA was successful in discriminating strains that carried glucose 1-6 linked to the galactose of O-polysaccharide backbone (serotype O1) from non-glucosylated strains. This was shown by O-antigen compositional analyses of the respective strains with mass spectrometry and capillary electrophoresis. The ELITA test worked rapidly in a microtiter plate format and was highly O-antigen specific. Moreover, TSP as probes could also detect glucosylated strains in flow cytometry and distinguish multiphasic cultures differing in their glucosylation state.
Conclusions
Tailspike proteins contain large binding sites with precisely defined specificities and are therefore promising tools to be included in serotyping procedures as rapid serotyping agents in addition to antibodies. In this study, 9NA and P22TSP as probes could specifically distinguish glucosylation phenotypes of Salmonella on microtiter plate assays and in flow cytometry. This opens the possibility for flow sorting of cell populations for subsequent genetic analyses or for monitoring phase variations during large scale O-antigen preparations necessary for vaccine production.
Background: Although nowaday it is broadly accepted that mitochondrial DNA (mtDNA) may undergo recombination, the frequency of such recombination remains controversial. Its estimation is not straightforward, as recombination under homoplasmy (i.e., among identical mt genomes) is likely to be overlooked. In species with tandem duplications of large mtDNA fragments the detection of recombination can be facilitated, as it can lead to gene conversion among duplicates. Although the mechanisms for concerted evolution in mtDNA are not fully understood yet, recombination rates have been estimated from "one per speciation event" down to 850 years or even "during every replication cycle".
Results: Here we present the first complete mt genome of the avian family Bucerotidae, i.e., that of two Philippine hornbills, Aceros waldeni and Penelopides panini. The mt genomes are characterized by a tandemly duplicated region encompassing part of cytochrome b, 3 tRNAs, NADH6, and the control region. The duplicated fragments are identical to each other except for a short section in domain I and for the length of repeat motifs in domain III of the control region. Due to the heteroplasmy with regard to the number of these repeat motifs, there is some size variation in both genomes; with around 21,657 bp (A. waldeni) and 22,737 bp (P. panini), they significantly exceed the hitherto longest known avian mt genomes, that of the albatrosses. We discovered concerted evolution between the duplicated fragments within individuals. The existence of differences between individuals in coding genes as well as in the control region, which are maintained between duplicates, indicates that recombination apparently occurs frequently, i. e., in every generation.
Conclusions: The homogenised duplicates are interspersed by a short fragment which shows no sign of recombination. We hypothesize that this region corresponds to the so-called Replication Fork Barrier (RFB), which has been described from the chicken mitochondrial genome. As this RFB is supposed to halt replication, it offers a potential mechanistic explanation for frequent recombination in mitochondrial genomes.
Background: The Visayan Tarictic Hornbill (Penelopides panini) and the Walden's Hornbill (Aceros waldeni) are two threatened hornbill species endemic to the western islands of the Visayas that constitute - between Luzon and Mindanao - the central island group of the Philippine archipelago. In order to evaluate their genetic diversity and to support efforts towards their conservation, we analyzed genetic variation in similar to 600 base pairs (bp) of the mitochondrial control region I and at 12-19 nuclear microsatellite loci. The sampling covered extant populations, still occurring only on two islands (P. panini: Panay and Negros, A. waldeni: only Panay), and it was augmented with museum specimens of extinct populations from neighboring islands. For comparison, their less endangered (= more abundant) sister taxa, the Luzon Tarictic Hornbill (P. manillae) from the Luzon and Polillo Islands and the Writhed Hornbill (A. leucocephalus) from Mindanao Island, were also included in the study. We reconstructed the population history of the two Penelopides species and assessed the genetic population structure of the remaining wild populations in all four species.
Results: Mitochondrial and nuclear data concordantly show a clear genetic separation according to the island of origin in both Penelopides species, but also unravel sporadic over-water movements between islands. We found evidence that deforestation in the last century influenced these migratory events. Both classes of markers and the comparison to museum specimens reveal a genetic diversity loss in both Visayan hornbill species, P. panini and A. waldeni, as compared to their more abundant relatives. This might have been caused by local extinction of genetically differentiated populations together with the dramatic decline in the abundance of the extant populations.
Conclusions: We demonstrated a loss in genetic diversity of P. panini and A. waldeni as compared to their sister taxa P. manillae and A. leucocephalus. Because of the low potential for gene flow and population exchange across islands, saving of the remaining birds of almost extinct local populations - be it in the wild or in captivity - is particularly important to preserve the species' genetic potential.
The all-female Amazon molly (Poecilia formosa) is the result of a hybridization of the Atlantic molly (P. mexicana) and the sailfin molly (P. latipinna) approximately 120,000 years ago. As a gynogenetic species, P. formosa needs to copulate with heterospecific males including males from one of its bisexual ancestral species. However, the sperm only triggers embryogenesis of the diploid eggs. The genetic information of the sperm donor typically will not contribute to the next generation of P. formosa. Hence, P. formosa possesses generally one allele from each of its ancestral species at any genetic locus. This raises the question whether both ancestral alleles are equally expressed in P. formosa. Allele-specific expression (ASE) has been previously assessed in various organisms, e.g., human and fish, and ASE was found to be important in the context of phenotypic variability and disease. In this study, we utilized Real-Time PCR techniques to estimate ASE of the androgen receptor alpha (arα) gene in several distinct tissues of Amazon mollies. We found an allelic bias favoring the maternal ancestor (P. mexicana) allele in ovarian tissue. This allelic bias was not observed in the gill or the brain tissue. Sequencing of the promoter regions of both alleles revealed an association between an Indel in a known CpG island and differential expression. Future studies may reveal whether our observed cis-regulatory divergence is caused by an ovary-specific trans-regulatory element, preferentially activating the allele of the maternal ancestor.
Background
Foot orthoses are usually assumed to be effective by optimizing mechanically dynamic rearfoot configuration. However, the effect from a foot orthosis on kinematics that has been demonstrated scientifically has only been marginal. The aim of this study was to examine the effect of different heights in medial arch-supported foot orthoses on rear foot motion during gait.
Methods
Nineteen asymptomatic runners (36±11years, 180±5cm, 79±10kg; 41±22km/week) participated in the study. Trials were recorded at 3.1 mph (5 km/h) on a treadmill. Athletes walked barefoot and with 4 different not customized medial arch-supported foot orthoses of various arch heights (N:0 mm, M:30 mm, H:35 mm, E:40mm). Six infrared cameras and the `Oxford Foot Model´ were used to capture motion. The average stride in each condition was calculated from 50 gait cycles per condition. Eversion excursion and internal tibia rotation were analyzed. Descriptive statistics included calculating the mean ± SD and 95% CIs. Group differences by condition were analyzed by one factor (foot orthoses) repeated measures ANOVA (α = 0.05).
Results
Eversion excursion revealed the lowest values for N and highest for H (B:4.6°±2.2°; 95% CI [3.1;6.2]/N:4.0°±1.7°; [2.9;5.2]/M:5.2°±2.6°; [3.6;6.8]/H:6.2°±3.3°; [4.0;8.5]/E:5.1°±3.5°; [2.8;7.5]) (p>0.05). Range of internal tibia rotation was lowest with orthosis H and highest with E (B:13.3°±3.2°; 95% CI [11.0;15.6]/N:14.5°±7.2°; [9.2;19.6]/M:13.8°±5.0°; [10.8;16.8]/H:12.3°±4.3°; [9.0;15.6]/E:14.9°±5.0°; [11.5;18.3]) (p>0.05). Differences between conditions were small and the intrasubject variation high.
Conclusion
Our results indicate that different arch support heights have no systematic effect on eversion excursion or the range of internal tibia rotation and therefore might not exert a crucial influence on rear foot alignment during gait.
Plants frequently have to weather both biotic and abiotic stressors, and have evolved sophisticated adaptation and defense mechanisms. In recent years, chromatin modifications, nucleosome positioning, and DNA methylation have been recognized as important components in these adaptations. Given their potential epigenetic nature, such modifications may provide a mechanistic basis for a stress memory, enabling plants to respond more efficiently to recurring stress or even to prepare their offspring for potential future assaults. In this review, we discuss both the involvement of chromatin in stress responses and the current evidence on somatic, intergenerational, and transgenerational stress memory.
Background
Recently, the incidence rate of back pain (BP) in adolescents has been reported at 21%. However, the development of BP in adolescent athletes is unclear. Hence, the purpose of this study was to examine the incidence of BP in young elite athletes in relation to gender and type of sport practiced.
Methods
Subjective BP was assessed in 321 elite adolescent athletes (m/f 57%/43%; 13.2 ± 1.4 years; 163.4 ± 11.4 cm; 52.6 ± 12.6 kg; 5.0 ± 2.6 training yrs; 7.6 ± 5.3 training h/week). Initially, all athletes were free of pain. The main outcome criterion was the incidence of back pain [%] analyzed in terms of pain development from the first measurement day (M1) to the second measurement day (M2) after 2.0 ± 1.0 year. Participants were classified into athletes who developed back pain (BPD) and athletes who did not develop back pain (nBPD). BP (acute or within the last 7 days) was assessed with a 5-step face scale (face 1–2 = no pain; face 3–5 = pain). BPD included all athletes who reported faces 1 and 2 at M1 and faces 3 to 5 at M2. nBPD were all athletes who reported face 1 or 2 at both M1 and M2. Data was analyzed descriptively. Additionally, a Chi2 test was used to analyze gender- and sport-specific differences (p = 0.05).
Results
Thirty-two athletes were categorized as BPD (10%). The gender difference was 5% (m/f: 12%/7%) but did not show statistical significance (p = 0.15). The incidence of BP ranged between 6 and 15% for the different sport categories. Game sports (15%) showed the highest, and explosive strength sports (6%) the lowest incidence. Anthropometrics or training characteristics did not significantly influence BPD (p = 0.14 gender to p = 0.90 sports; r2 = 0.0825).
Conclusions
BP incidence was lower in adolescent athletes compared to young non-athletes and even to the general adult population. Consequently, it can be concluded that high-performance sports do not lead to an additional increase in back pain incidence during early adolescence. Nevertheless, back pain prevention programs should be implemented into daily training routines for sport categories identified as showing high incidence rates.
In the context of back pain, great emphasis has been placed on the importance of trunk stability, especially in situations requiring compensation of repetitive, intense loading induced during high-performance activities, e.g., jumping or landing. This study aims to evaluate trunk muscle activity during drop jump in adolescent athletes with back pain (BP) compared to athletes without back pain (NBP). Eleven adolescent athletes suffering back pain (BP: m/f: n = 4/7; 15.9 ± 1.3 y; 176 ± 11 cm; 68 ± 11 kg; 12.4 ± 10.5 h/we training) and 11 matched athletes without back pain (NBP: m/f: n = 4/7; 15.5 ± 1.3 y; 174 ± 7 cm; 67 ± 8 kg; 14.9 ± 9.5 h/we training) were evaluated. Subjects conducted 3 drop jumps onto a force plate (ground reaction force). Bilateral 12-lead SEMG (surface Electromyography) was applied to assess trunk muscle activity. Ground contact time [ms], maximum vertical jump force [N], jump time [ms] and the jump performance index [m/s] were calculated for drop jumps. SEMG amplitudes (RMS: root mean square [%]) for all 12 single muscles were normalized to MIVC (maximum isometric voluntary contraction) and analyzed in 4 time windows (100 ms pre- and 200 ms post-initial ground contact, 100 ms pre- and 200 ms post-landing) as outcome variables. In addition, muscles were grouped and analyzed in ventral and dorsal muscles, as well as straight and transverse trunk muscles. Drop jump ground reaction force variables did not differ between NBP and BP (p > 0.05). Mm obliquus externus and internus abdominis presented higher SEMG amplitudes (1.3–1.9-fold) for BP (p < 0.05). Mm rectus abdominis, erector spinae thoracic/lumbar and latissimus dorsi did not differ (p > 0.05). The muscle group analysis over the whole jumping cycle showed statistically significantly higher SEMG amplitudes for BP in the ventral (p = 0.031) and transverse muscles (p = 0.020) compared to NBP. Higher activity of transverse, but not straight, trunk muscles might indicate a specific compensation strategy to support trunk stability in athletes with back pain during drop jumps. Therefore, exercises favoring the transverse trunk muscles could be recommended for back pain treatment.
This study aimed to determine the relative and absolute reliability of ultrasound (US) measurements of the thickness and echogenicity of the plantar fascia (PF) at different measurement stations along its length using a standardized protocol. Twelve healthy subjects (24 feet) were enrolled. The PF was imaged in the longitudinal plane. Subjects were assessed twice to evaluate the intra-rater reliability. A quantitative evaluation of the thickness and echogenicity of the plantar fascia was performed using Image J, a digital image analysis and viewer software. A sonography evaluation of the thickness and echogenicity of the PF showed a high relative reliability with an Intra class correlation coefficient of 0.88 at all measurement stations. However, the measurement stations for both the PF thickness and echogenicity which showed the highest intraclass correlation coefficient (ICCs) did not have the highest absolute reliability. Compared to other measurement stations, measuring the PF thickness at 3 cm distal and the echogenicity at a region of interest 1 cm to 2 cm distal from its insertion at the medial calcaneal tubercle showed the highest absolute reliability with the least systematic bias and random error. Also, the reliability was higher using a mean of three measurements compared to one measurement. To reduce discrepancies in the interpretation of the thickness and echogenicity measurements of the PF, the absolute reliability of the different measurement stations should be considered in clinical practice and research rather than the relative reliability with the ICC.
Decades of research have demonstrated that physical stress (PS) stimulates bone remodeling and affects bone structure and function through complex mechanotransduction mechanisms. Recent research has laid ground to the hypothesis that mental stress (MS) also influences bone biology, eventually leading to osteoporosis and increased bone fracture risk. These effects are likely exerted by modulation of hypothalamic–pituitary–adrenal axis activity, resulting in an altered release of growth hormones, glucocorticoids and cytokines, as demonstrated in human and animal studies. Furthermore, molecular cross talk between mental and PS is thought to exist, with either synergistic or preventative effects on bone disease progression depending on the characteristics of the applied stressor. This mini review will explain the emerging concept of MS as an important player in bone adaptation and its potential cross talk with PS by summarizing the current state of knowledge, highlighting newly evolving notions (such as intergenerational transmission of stress and its epigenetic modifications affecting bone) and proposing new research directions.
Objectives: Chronic back pain (CBP) can lead to disability and burden. In addition to its medical causes, its development is influenced by psychosocial risk factors, the so-called flag factors, which are categorized and integrated into many treatment guidelines. Currently, most studies investigate single flag factors, which limit the estimation of individual factor significance in the development of chronic pain. Furthermore, factors concerning patients’ lifestyle, biography and treatment history are often neglected. Therefore, the objectives of the present study are to identify commonly neglected factors of CBP and integrate them into an analysis model comparing their significance with established flag factors.
Methods: A total of 24 patients and therapists were cross-sectionally interviewed to identify commonly neglected factors of CBP. Subsequently, the impact of these factors was surveyed in a longitudinal study. In two rehabilitation clinics, CBP patients (n = 145) were examined before and 6 months after a 3-week inpatient rehabilitation. Outcome variables, chronification factor pain experience (CF-PE) and chronification factor disability (CF-D), were ascertained with confirmatory factor analysis (CFA) of standardized questionnaires. Predictors were evaluated using stepwise calculations of simple and multiple regression models.
Results: Through interviews, medical history, iatrogenic factors, poor compliance, critical life events (LEs), social support (SS) type and effort–reward were identified as commonly neglected factors. However, only the final three held significance in comparison to established factors such as depression and pain-related cognitions. Longitudinally, lifestyle factors found to influence future pain were initial pain, physically demanding work, nicotine consumption, gender and rehabilitation clinic. LEs were unexpectedly found to be a strong predictor of future pain, as were the protective factors, reward at work and perceived SS.
Discussion: These findings shed insight regarding often overlooked factors in the development of CBP, suggesting that more detailed operationalization and superordinate frameworks would be beneficial to further research.
Conclusion: In particular, LEs should be taken into account in future research. Protective factors should be integrated in therapeutic settings.
A particular form of social pain is invalidation. Therefore, this study (a) investigates whether patients with chronic low back pain experience invalidation, (b) if it has an influence on their pain, and (c) explores whether various social sources (e.g. partner and work) influence physical pain differentially. A total of 92 patients completed questionnaires, and for analysis, Pearson's correlation coefficients and hierarchical linear regression analyses were conducted. They indicated a significant association between discounting and disability due to pain (respective =.29, p>.05). Especially, discounting by partner was linked to higher disability (=.28, p>.05).
BACKGROUND: Reduced left ventricular ejection fraction (LVEF) ≤30% is the most powerful prognostic indicator for sudden cardiac death (SCD) in patients after myocardial infarction (MI), but there are little data about long-term changes of LVEF after revascularization and the following implantation of a cardioverter defibrillator (ICD).
METHODS: We performed a retrospective analysis of 277 patients with reduced LVEF at least 1month after MI and complete revascularization. Patients (median time post-MI 23.4months; 74.3% after PCI, 25.7% after CABG were assigned either to group 1 (LVEF<30%) or group 2 (LVEF 30-40%). Biplane echocardiography was redone after a mean follow-up of 441±220days.
RESULTS: LVEF increased significantly in both two groups (group 1: 26.2±4.8% to 32.4±8.5%; p<0.001; group 2: 38.2±2.5% to 44.4±9.6%; p<0.001). However, statistical analysis of first and second LVEF measurement by means of a LOWESS regression and with an appropriate correction of the regression towards the mean effect revealed only a moderate increase of the mean LVEF from 35 to 37% (p<0.001) with a large interindividual variation.
CONCLUSIONS: The impact of early revascularization on LVEF appears to be low in the majority of post-MI heart failure patients. Owing to the high variability, a single measurement may not be reliable enough to justify a decision on ICD indication.
The Star Excursion Balance Test (SEBT) is effective in measuring dynamic postural control (DPC). This research aimed to determine whether DPC measured by the SEBT in young athletes (YA) with back pain (BP) is different from those without BP (NBP). 53 BP YA and 53 NBP YA matched for age, height, weight, training years, training sessions/week and training minutes/session were studied. Participants performed 4 practice trials after which 3 measurements in the anterior, posteromedial and posterolateral SEBT reach directions were recorded. Normalized reach distance was analyzed using the mean of all 3 measurements. There was no statistical significant difference (p > 0.05) between the reach distance of BP (87.2 ± 5.3, 82.4 ± 8.2, 78.7 ± 8.1) and NBP (87.8 ± 5.6, 82.4 ± 8.0, 80.0 ± 8.8) in the anterior, posteromedial and posterolateral directions respectively. DPC in YA with BP, as assessed by the SEBT, was not different from NBP YA.
This review analyzes the potential role and long-term effects of field perennial polycultures (mixtures) in agricultural systems, with the aim of reducing the trade-offs between provisioning and regulating ecosystem services. First, crop rotations are identified as a suitable tool for the assessment of the long-term effects of perennial polycultures on ecosystem services, which are not visible at the single-crop level. Second, the ability of perennial polycultures to support ecosystem services when used in crop rotations is quantified through eight agricultural ecosystem services. Legume-grass mixtures and wildflower mixtures are used as examples of perennial polycultures, and compared with silage maize as a typical crop for biomass production. Perennial polycultures enhance soil fertility, soil protection, climate regulation, pollination, pest and weed control, and landscape aesthetics compared with maize. They also score lower for biomass production compared with maize, which confirms the trade-off between provisioning and regulating ecosystem services. However, the additional positive factors provided by perennial polycultures, such as reduced costs for mineral fertilizer, pesticides, and soil tillage, and a significant preceding crop effect that increases the yields of subsequent crops, should be taken into account. However, a full assessment of agricultural ecosystem services requires a more holistic analysis that is beyond the capabilities of current frameworks.
Modifications of transfer RNA (tRNA) have been shown to play critical roles in the biogenesis, metabolism, structural stability and function of RNA molecules, and the specific modifications of nucleobases with sulfur atoms in tRNA are present in pro- and eukaryotes. Here, especially the thiomodifications xm(5)s(2)U at the wobble position 34 in tRNAs for Lys, Gln and Glu, were suggested to have an important role during the translation process by ensuring accurate deciphering of the genetic code and by stabilization of the tRNA structure. The trafficking and delivery of sulfur nucleosides is a complex process carried out by sulfur relay systems involving numerous proteins, which not only deliver sulfur to the specific tRNAs but also to other sulfur-containing molecules including iron-sulfur clusters, thiamin, biotin, lipoic acid and molybdopterin (MPT). Among the biosynthesis of these sulfur-containing molecules, the biosynthesis of the molybdenum cofactor (Moco) and the synthesis of thio-modified tRNAs in particular show a surprising link by sharing protein components for sulfur mobilization in pro- and eukaryotes.
Sequelae of prematurity triggered by oxidative stress and free radical-mediated tissue damage have coined the term "oxygen radical disease of prematurity". Caffeine, a potent free radical scavenger and adenosine receptor antagonist, reduces rates of brain damage in preterm infants. In the present study, we investigated the effects of caffeine on oxidative stress markers, anti-oxidative response, inflammation, redox-sensitive transcription factors, apoptosis, and extracellular matrix following the induction of hyperoxia in neonatal rats. The brain of a rat pups at postnatal Day 6 (P6) corresponds to that of a human fetal brain at 28-32 weeks gestation and the neonatal rat is an ideal model in which to investigate effects of oxidative stress and neuroprotection of caffeine on the developing brain. Six-day-old Wistar rats were pre-treated with caffeine and exposed to 80% oxygen for 24 and 48 h. Caffeine reduced oxidative stress marker (heme oxygenase-1, lipid peroxidation, hydrogen peroxide, and glutamate-cysteine ligase catalytic subunit (GCLC)), promoted anti-oxidative response (superoxide dismutase, peroxiredoxin 1, and sulfiredoxin 1), down-regulated pro-inflammatory cytokines, modulated redox-sensitive transcription factor expression (Nrf2/Keap1, and NF kappa B), reduced pro-apoptotic effectors (poly (ADP-ribose) polymerase-1 (PARP-1), apoptosis inducing factor (AIF), and caspase-3), and diminished extracellular matrix degeneration (matrix metalloproteinases (MMP) 2, and inhibitor of metalloproteinase (TIMP) 1/2). Our study affirms that caffeine is a pleiotropic neuroprotective drug in the developing brain due to its anti-oxidant, anti-inflammatory, and anti-apoptotic properties.
Molecularly imprinted polymers (MIPs) have the potential to complement antibodies in bioanalysis, are more stable under harsh conditions, and are potentially cheaper to produce. However, the affinity and especially the selectivity of MIPs are in general lower than those of their biological pendants. Enzymes are useful tools for the preparation of MIPs for both low and high-molecular weight targets: As a green alternative to the well-established methods of chemical polymerization, enzyme-initiated polymerization has been introduced and the removal of protein templates by proteases has been successfully applied. Furthermore, MIPs have been coupled with enzymes in order to enhance the analytical performance of biomimetic sensors: Enzymes have been used in MIP-sensors as tracers for the generation and amplification of the measuring signal. In addition, enzymatic pretreatment of an analyte can extend the analyte spectrum and eliminate interferences.
In assessing adolescent behavior difficulties, parents, teachers, and the adolescents themselves are key informants. However, substantial disagreement has been found between informants. Specifically, children with attention-deficit/hyperactivity disorder (ADHD) tend to overestimate their competencies, also known as “positive (illusionary) bias.” This study compared parent, teacher, and adolescent ratings of ADHD and other behavioral symptoms in a sample of 114 adolescents with ADHD. Further, the effect of cross-informant disagreement (CID) on treatment outcomes was investigated in a subsample of 54 adolescents who had undergone a training and coaching intervention. Overall, there was moderate agreement among informants. Parent and adolescent ratings were more strongly correlated with each other than with teacher ratings. The strongest discrepancy was found between teacher and adolescent ratings on prosocial behavior. This discrepancy explained 12% of the variance in parent-rated ADHD symptom severity after the intervention. The treatment was less effective in participants with high teacher-adolescent disagreement on prosocial behavior (d = 0.41) than with low disagreement (d = 0.98). These findings suggest that professionals working with adolescents with ADHD should consider multiple sources of information before initiating treatment and pay attention to cross-informant disagreements because these may indicate a risk of diminished treatment effects.
In October 2016, following a campaign led by Labour Peer Lord
Alfred Dubs, the first child asylum-seekers allowed entry to the UK
under new legislation (the ‘Dubs amendment’) arrived in England.
Their arrival was captured by a heavy media presence, and very
quickly doubts were raised by right-wing tabloids and politicians
about their age. In this article, I explore the arguments
underpinning the Dubs campaign and the media coverage of
the children’s arrival as a starting point for interrogating
representational practices around children who seek asylum. I
illustrate how the campaign was premised on a universal politics
of childhood that inadvertently laid down the terms on which
these children would be given protection, namely their innocence.
The universality of childhood fuels public sympathy for child
asylum-seekers, underlies the ‘child first, migrant second’
approach advocated by humanitarian organisations, and it was a
key argument in the ‘Dubs amendment’. Yet the campaign
highlights how representations of child asylum-seekers rely on
codes that operate to identify ‘unchildlike’ children. As I show, in
the context of the criminalisation of undocumented migrants‘,
childhood is no longer a stable category which guarantees
protection, but is subject to scrutiny and suspicion and can,
ultimately, be disproved.
Reaching the Sustainable Development Goals requires a fundamental socio-economic transformation accompanied by substantial investment in low-carbon infrastructure. Such a sustainability transition represents a non-marginal change, driven by behavioral factors and systemic interactions. However, typical economic models used to assess a sustainability transition focus on marginal changes around a local optimum, whichby constructionlead to negative effects. Thus, these models do not allow evaluating a sustainability transition that might have substantial positive effects. This paper examines which mechanisms need to be included in a standard computable general equilibrium model to overcome these limitations and to give a more comprehensive view of the effects of climate change mitigation. Simulation results show that, given an ambitious greenhouse gas emission constraint and a price of carbon, positive economic effects are possible if (1) technical progress results (partly) endogenously from the model and (2) a policy intervention triggering an increase of investment is introduced. Additionally, if (3) the investment behavior of firms is influenced by their sales expectations, the effects are amplified. The results provide suggestions for policy-makers, because the outcome indicates that investment-oriented climate policies can lead to more desirable outcomes in economic, social and environmental terms.
Background: Gestational diabetes mellitus (GDM) is associated with adverse pregnancy outcomes. It is known that GDM is associated with an altered placental function and changes in placental gene regulation. More recent studies demonstrated an involvement of epigenetic mechanisms. So far, the focus regarding placental epigenetic changes in GDM was set on gene-specific DNA methylation analyses. Studies that robustly investigated placental global DNA methylation are lacking. However, several studies showed that tissue-specific alterations in global DNA methylation are independently associated with type 2 diabetes. Thus, the aim of this study was to characterize global placental DNA methylation by robustly measuring placental DNA 5-methylcytosine (5mC) content and to examine whether differences in placental global DNA methylation are associated with GDM.
Methods: Global DNA methylation was quantified by the current gold standard method, LC-MS/MS. In total, 1030 placental samples were analyzed in this single-center birth cohort study.
Results: Mothers with GDM displayed a significantly increased global placental DNA methylation (3.22 ± 0.63 vs. 3.00 ± 0.46 %; p = 0.013; ±SD). Bivariate logistic regression showed a highly significant positive correlation between global placental DNA methylation and the presence of GDM (p = 0.0009). Quintile stratification according to placental DNA 5mC levels revealed that the frequency of GDM was evenly distributed in quintiles 1–4 (2.9–5.3 %), whereas the frequency in the fifth quintile was significantly higher (10.7 %; p = 0.003). Bivariate logistic models adjusted for maternal age, BMI, ethnicity, recurrent miscarriages, and familiar diabetes predisposition clearly demonstrated an independent association between global placental DNA hypermethylation and GDM. Furthermore, an ANCOVA model considering known predictors of DNA methylation substantiated an independent association between GDM and placental DNA methylation.
Conclusions: This is the first study that employed a robust quantitative assessment of placental global DNA methylation in over a thousand placental samples. The study provides large scale evidence that placental global DNA hypermethylation is associated with GDM, independent of established risk factors.
The ecological benefits of polyploidy are intensely debated. Some authors argue that plants with duplicated chromosome sets (polyploids) are more stress-resistant and superior colonizers and may thus outnumber their low ploidy conspecifics in more extreme habitats. Brachypodium distachyon (sensu lato), for example, a common annual grass in Israel and the entire Mediterranean basin, comprises three cytotypes of differing chromosome numbers that were recently proposed as distinct species. It was suggested that increased aridity increases the occurrence of its polyploid cytotype. Here, we tested at two spatial scales whether polyploid plants of B. distachyon s.l. are more frequently found in drier habitats in Israel. We collected a total of 430 specimens (i) along a largescale climatic gradient with 15 thoroughly selected sites (spanning 114–954 mm annual rainfall), and (ii) from corresponding Northern (more mesic) and Southern (more arid) hill slopes to assess the micro-climatic difference between contrasting exposures. Cytotypes were then determined via flow cytometry. Polyploid plants comprised 90% of all specimens and their proportion ranged between 0% and 100% per site. However, this proportion was not correlated with aridity along the large-scale gradient, nor were polyploids more frequently found on Southern exposures. Our results show for both spatial scales that increasing aridity is not the principal driver for the distribution of polyploids in B. distachyon s.l. in Israel. Notably, though, diploid plants were restricted essentially to four intermediate sites, while polyploids dominated the most arid and the most mesic sites. This, to some degree, clustered pattern suggests that the distribution of cytotypes is not entirely random and calls for future studies to assess further potential drivers.
VS30, slope, H800 and f0
(2017)
The aim of this paper is to investigate the ability of various site-condition proxies (SCPs) to reduce ground-motion aleatory variability and evaluate how SCPs capture nonlinearity site effects. The SCPs used here are time-averaged shear-wave velocity in the top 30 m (VS30), the topographical slope (slope), the fundamental resonance frequency (f0) and the depth beyond which Vs exceeds 800 m/s (H800). We considered first the performance of each SCP taken alone and then the combined performance of the 6 SCP pairs [VS30–f0], [VS30–H800], [f0–slope], [H800–slope], [VS30–slope] and [f0–H800]. This analysis is performed using a neural network approach including a random effect applied on a KiK-net subset for derivation of ground-motion prediction equations setting the relationship between various ground-motion parameters such as peak ground acceleration, peak ground velocity and pseudo-spectral acceleration PSA (T), and Mw, RJB, focal depth and SCPs. While the choice of SCP is found to have almost no impact on the median groundmotion prediction, it does impact the level of aleatory uncertainty. VS30 is found to perform the best of single proxies
at short periods (T < 0.6 s), while f0 and H800 perform better at longer periods; considering SCP pairs leads to significant improvements, with particular emphasis on [VS30–H800] and [f0–slope] pairs. The results also indicate significant nonlinearity on the site terms for soft sites and that the most relevant loading parameter for characterising nonlinear site response is the “stiff” spectral ordinate at the considered period.
School adjustment determines long-term adjustment in society. Yet, immigrant youth do better in some countries than in others. Drawing on acculturation research (Berry, 1997; Ward, 2001) and self-determination theory (Ryan and Deci, 2000), we investigated indirect effects of adolescent immigrants’ acculturation orientations on school adjustment (school-related attitudes, truancy, and mathematics achievement) through school belonging. Analyses were based on data from the Programme for International Student Assessment from six European countries, which were combined into three clusters based on their migrant integration and multicultural policies: Those with the most supportive policies (Belgium and Finland), those with moderately supportive policies (Italy and Portugal), and those with the most unsupportive policies (Denmark and Slovenia). In a multigroup path model, we confirmed most associations. As expected, mainstream orientation predicted higher belonging and better outcomes in all clusters, whereas the added value of students’ ethnic orientation was only observed in some clusters. Results are discussed in terms of differences in acculturative climate and policies between countries of settlement.
Schools are a major context for academic and socio-emotional development, but
also an important acculturative context. This is notably the case in adolescence,
which is a critical period for the development of a social and ethnic identity, as
well as moral reasoning and intergroup attitudes. How schools approach cultural
diversity issues is therefore likely to affect these developmental and acculturative
processes and adaptation outcomes. In the present article, the manifestation
and effects of the most prominent approaches to cultural diversity, namely
those guided by a perspective of equality and inclusion, and those guided by
a perspective of cultural pluralism, are reviewed and compared in the context
of multi-ethnic schools. The aim is to explore when and how the potential of
cultural diversity can best flourish, enhancing the academic and socio-emotional
development of culturally diverse students.
Across currents
(2017)
The purpose of the project was to develop the Discounting Inventory (DI), a measure of individual differences in delay, probability, effort, and social discounting, all related to behavioral impulsivity. Over 400 items relating to four types of discounting were generated. Next, a study followed by a series of psychometric analyses of data obtained from a group of 2843 individuals was conducted. Principal Component Analysis yielded a four-factor structure of data, reflecting the four types of discounting. The results of Confirmatory Factor Analysis showed good fit of the four-factor model to data. Through several iterations of retaining and deleting items on the basis of their component loadings, item intercorrelations, and contribution to coefficient alphas, the total number of items was reduced to 48. The final 48-item version of the inventory has satisfactory psychometric characteristics, including Cronbach’s alpha and test–retest stability. In addition, significant correlations were observed between the DI and traditional discounting instruments, suggesting that the DI measures a construct similar to the behavioral discounting process. The development of the tool was based on the assumption that discounting is a personality trait. However, the present data suggest that discounting may reflect more a state than trait function.
Cancer cachexia, of which the most notable symptom is severe and rapid weight loss, is present in the majority of patients with advanced cancer. Inflammatory mediators play an important role in the development of cachexia, envisaged as a chronic inflammatory syndrome. The white adipose tissue (WAT) is one of the first compartments affected in cancer cachexia and suffers a high rate of lipolysis. It secretes several cytokines capable of directly regulating intermediate metabolism. A common pathway in the regulation of the expression of pro-inflammatory cytokines in WAT is the activation of the nuclear transcription factor kappa-B (NF-κB). We have examined the gene expression of the subunits NF-κBp65 and NF-κBp50, as well as NF-κBp65 and NF-κBp50 binding, the gene expression of pro-inflammatory mediators under NF-κB control (IL-1β, IL-6, INF-γ, TNF-α, MCP-1), and its inhibitory protein, nuclear factor of kappa light polypeptide gene enhancer in B-cells inhibitor, alpha (IκB-α). The observational study involved 35 patients (control group, n = 12 and cancer group, n = 23, further divided into cachectic and non-cachectic). NF-κBp65 and its target genes expression (TNF-α, IL-1β, MCP-1 and IκB-α) were significantly higher in cachectic cancer patients. Moreover, NF-κBp65 gene expression correlated positively with the expression of its target genes. The results strongly suggest that the NF-κB pathway plays a role in the promotion of WAT inflammation during cachexia.
The transition from hunting and gathering to farming involved profound cultural and technological changes. In Western and Central Europe, these changes occurred rapidly and synchronously after the arrival of early farmers of Anatolian origin [1-3], who largely replaced the local Mesolithic hunter-gatherers [1, 4-6]. Further east, in the Baltic region, the transition was gradual, with little or no genetic input from incoming farmers [7]. Here we use ancient DNA to investigate the relationship between hunter-gatherers and farmers in the Lower Danube basin, a geographically intermediate area that is characterized by a rapid Neolithic transition but also by the presence of archaeological evidence that points to cultural exchange, and thus possible admixture, between hunter-gatherers and farmers. We recovered four human paleogenomes (1.13 to 4.13 coverage) from Romania spanning a time transect between 8.8 thousand years ago (kya) and 5.4 kya and supplemented them with two Mesolithic genomes (1.73- and 5.33) from Spain to provide further context on the genetic background of Mesolithic Europe. Our results show major Western hunter-gatherer (WHG) ancestry in a Romanian Eneolithic sample with a minor, but sizeable, contribution from Anatolian farmers, suggesting multiple admixture events between hunter-gatherers and farmers. Dietary stableisotope analysis of this sample suggests a mixed terrestrial/ aquatic diet. Our results provide support for complex interactions among hunter-gatherers and farmers in the Danube basin, demonstrating that in some regions, demic and cultural diffusion were not mutually exclusive, but merely the ends of a continuum for the process of Neolithization.
Palaeogenomes of Eurasian straight-tusked elephants challenge the current view of elephant evolution
(2017)
The straight-tusked elephants Palaeoloxodon spp. were widespread across Eurasia during the Pleistocene. Phylogenetic reconstructions using morphological traits have grouped them with Asian elephants (Elephas maximus), and many paleontologists place Palaeoloxodon within Elephas. Here, we report the recovery of full mitochondrial genomes from four and partial nuclear genomes from two P. antiquus fossils. These fossils were collected at two sites in Germany, Neumark-Nord and Weimar-Ehringsdorf, and likely date to interglacial periods similar to 120 and similar to 244 thousand years ago, respectively. Unexpectedly, nuclear and mitochondrial DNA analyses suggest that P. antiquus was a close relative of extant African forest elephants (Loxodonta cyclotis). Species previously referred to Palaeoloxodon are thus most parsimoniously explained as having diverged from the lineage of Loxodonta, indicating that Loxodonta has not been constrained to Africa. Our results demonstrate that the current picture of elephant evolution is in need of substantial revision.
Lithospheric plates move over the low viscosity asthenosphere balancing several forces. The driving forces include basal shear stress exerted by mantle convection and plate boundary forces such as slab pull and ridge push, whereas the resisting forces include inter-plate friction, trench resistance, and cratonic root resistance. These generate plate motions, the lithospheric stress field and dynamic topography which are observed with different geophysical methods. The orientation and tectonic regime of the observed crustal/lithospheric stress field further contribute to our knowledge of different deformation processes occurring within the Earth's crust and lithosphere. Using numerical models previous studies were able to identify major forces generating stresses in the crust and lithosphere which also contribute to the formation of topography as well as driving lithospheric plates. They showed that the first-order stress pattern explaining about 80\,\% of the stress field originates from a balance of forces acting at the base of the moving lithospheric plates due to convective flow in the underlying mantle. The remaining second-order stress pattern is due to lateral density variations in the crust and lithosphere in regions of pronounced topography and high gravitational potential, such as the Himalayas and mid-ocean ridges. By linking global lithosphere dynamics to deep mantle flow this study seeks to evaluate the influence of shallow and deep density heterogenities on plate motions, lithospheric stress field and dynamic topography using the geoid as a major constraint for mantle rheology. We use the global 3D lithosphere-asthenosphere model SLIM3D with visco-elasto-plastic rheology coupled at 300 km depth to a spectral model of mantle flow. The complexity of the lithosphere-asthenosphere component allows for the simulation of power-law rheology with creep parameters accounting for both diffusion and dislocation creep within the uppermost 300 km.
First we investigate the influence of intra-plate friction and asthenospheric viscosity on present-day plate motions. Previous modelling studies have suggested that small friction coefficients (µ < 0.1, yield stress ~ 100 MPa) can lead to plate tectonics in models of mantle convection. Here we show that, in order to match present-day plate motions and net rotation, the frictional parameter must be less than 0.05. We are able to obtain a good fit with the magnitude and orientation of observed plate velocities (NUVEL-1A) in a no-net-rotation (NNR) reference frame with µ < 0.04 and minimum asthenosphere viscosity ~ 5*10e19 Pas to 10e20 Pas. Our estimates of net rotation (NR) of the lithosphere suggest that amplitudes ~ 0.1-0.2 °/Ma, similar to most observation-based estimates, can be obtained with asthenosphere viscosity cutoff values of ~ 10e19 Pas to 5*10e19 Pas and friction coefficient µ < 0.05.
The second part of the study investigates further constraints on shallow and deep mantle heterogeneities causing plate motion by predicting lithosphere stress field and topography and validating with observations. Lithosphere stresses and dynamic topography are computed using the modelling setup and rheological parameters for prescribed plate motions. We validate our results with the World Stress Map 2016 (WSM2016) and the observed residual topography. Here we tested a number of upper mantle thermal-density structures. The one used to calculate plate motions is considered the reference thermal-density structure. This model is derived from a heat flow model combined with a sea floor age model. In addition we used three different thermal-density structures derived from global S-wave velocity models to show the influence of lateral density heterogeneities in the upper 300 km on model predictions. A large portion of the total dynamic force generating stresses in the crust/lithosphere has its origin in the deep mantle, while topography is largely influenced by shallow heterogeneities. For example, there is hardly any difference between the stress orientation patterns predicted with and without consideration of the heterogeneities in the upper mantle density structure across North America, Australia, and North Africa. However, the crust is dominant in areas of high altitude for the stress orientation compared to the all deep mantle contribution.
This study explores the sensitivity of all the considered surface observables with regards to model parameters providing insights into the influence of the asthenosphere and plate boundary rheology on plate motion as we test various thermal-density structures to predict stresses and topography.
According to the classical plume hypothesis, mantle plumes are localized upwellings of hot, buoyant material in the Earth’s mantle. They have a typical mushroom shape, consisting of a large plume head, which is associated with the formation of voluminous flood basalts (a Large
Igneous Province) and a narrow plume tail, which generates a linear, age-progressive chain of volcanic edifices (a hotspot track) as the tectonic plate migrates over the relatively stationary plume. Both plume heads and tails reshape large areas of the Earth’s surface over many tens of millions of years.
However, not every plume has left an exemplary record that supports the classical hypothesis. The main objective of this thesis is therefore to study how specific hotspots have created the crustal thickness pattern attributed to their volcanic activities. Using regional geodynamic
models, the main chapters of this thesis address the challenge of deciphering the three individual (and increasingly complex) Réunion, Iceland, and Kerguelen hotspot histories, especially focussing on the interactions between the respective plume and nearby spreading ridges.
For this purpose, the mantle convection code ASPECT is used to set up three-dimensional numerical models, which consider the specific local surroundings of each plume by prescribing time-dependent boundary conditions for temperature and mantle flow. Combining reconstructed plate boundaries and plate motions, large-scale global flow velocities and an inhomogeneous lithosphere thickness distribution together with a dehydration rheology represents a novel setup for regional convection models.
The model results show the crustal thickness pattern produced by the plume, which is compared to present-day topographic structures, crustal thickness estimates and age determinations of volcanic provinces associated with hotspot activity. Altogether, the model results agree well
with surface observations. Moreover, the dynamic development of the plumes in the models provide explanations for the generation of smaller, yet characteristic volcanic features that were previously unexplained. Considering the present-day state of a model as a prediction for the
current temperature distribution in the mantle, it cannot only be compared to observations on the surface, but also to structures in the Earth’s interior as imaged by seismic tomography.
More precisely, in the case of the Réunion hotspot, the model demonstrates how the distinctive gap between the Maldives and Chagos is generated due to the combination of the ridge geometry and plume-ridge interaction. Further, the Rodrigues Ridge is formed as the surface expression
of a long-distance sublithospheric flow channel between the upwelling plume and the closest ridge segment, confirming the long-standing hypothesis of Morgan (1978) for the first time in a dynamic context. The Réunion plume has been studied in connection with the seismological
RHUM-RUM project, which has recently provided new seismic tomography images that yield an excellent match with the geodynamic model.
Regarding the Iceland plume, the numerical model shows how plume material may have accumulated in an east-west trending corridor of thin lithosphere across Greenland and resulted in simultaneous melt generation west and east of Greenland. This provides an explanation for the
extremely widespread volcanic material attributed to magma production of the Iceland hotspot and demonstrates that the model setup is also able to explain more complicated hotspot histories. The Iceland model results also agree well with newly derived seismic tomographic images.
The Kerguelen hotspot has an extremely complex history and previous studies concluded that the plume might be dismembered or influenced by solitary waves in its conduit to produce the reconstructed variable melt production rate. The geodynamic model, however, shows that a constant plume influx can result in a variable magma production rate if the plume interacts with nearby mid-ocean ridges. Moreover, the Ninetyeast Ridge in the model is created by on-ridge activities, while the Kerguelen plume was located beneath the Australian plate. This is also a contrast to earlier studies, which described the Ninetyeast Ridge as the result of the Indian plate passing over the plume. Furthermore, the Amsterdam-Saint Paul Plateau in the model is the result of plume material flowing from the upwelling toward the Southeast Indian Ridge, whereas previous geochemical studies attributed that volcanic province to a separate deep plume.
In summary, the three case studies presented in this thesis consistently highlight the importance of plume-ridge interaction in order to reconstruct the overall volcanic hotspot record as well as specific smaller features attributed to a certain hotspot. They also demonstrate that it is not necessary to attribute highly complicated properties to a specific plume in order to account for complex observations. Thus, this thesis contributes to the general understanding of plume dynamics and extends the very specific knowledge about the Réunion, Iceland, and Kerguelen mantle plumes.
With recent advances in the area of information extraction, automatically extracting structured information from a vast amount of unstructured textual data becomes an important task, which is infeasible for humans to capture all information manually. Named entities (e.g., persons, organizations, and locations), which are crucial components in texts, are usually the subjects of structured information from textual documents. Therefore, the task of named entity mining receives much attention. It consists of three major subtasks, which are named entity recognition, named entity linking, and relation extraction.
These three tasks build up an entire pipeline of a named entity mining system, where each of them has its challenges and can be employed for further applications. As a fundamental task in the natural language processing domain, studies on named entity recognition have a long history, and many existing approaches produce reliable results. The task is aiming to extract mentions of named entities in text and identify their types. Named entity linking recently received much attention with the development of knowledge bases that contain rich information about entities. The goal is to disambiguate mentions of named entities and to link them to the corresponding entries in a knowledge base. Relation extraction, as the final step of named entity mining, is a highly challenging task, which is to extract semantic relations between named entities, e.g., the ownership relation between two companies.
In this thesis, we review the state-of-the-art of named entity mining domain in detail, including valuable features, techniques, evaluation methodologies, and so on. Furthermore, we present two of our approaches that focus on the named entity linking and relation extraction tasks separately.
To solve the named entity linking task, we propose the entity linking technique, BEL, which operates on a textual range of relevant terms and aggregates decisions from an ensemble of simple classifiers. Each of the classifiers operates on a randomly sampled subset of the above range. In extensive experiments on hand-labeled and benchmark datasets, our approach outperformed state-of-the-art entity linking techniques, both in terms of quality and efficiency.
For the task of relation extraction, we focus on extracting a specific group of difficult relation types, business relations between companies. These relations can be used to gain valuable insight into the interactions between companies and perform complex analytics, such as predicting risk or valuating companies. Our semi-supervised strategy can extract business relations between companies based on only a few user-provided seed company pairs. By doing so, we also provide a solution for the problem of determining the direction of asymmetric relations, such as the ownership_of relation. We improve the reliability of the extraction process by using a holistic pattern identification method, which classifies the generated extraction patterns. Our experiments show that we can accurately and reliably extract new entity pairs occurring in the target relation by using as few as five labeled seed pairs.
The article explores Europeanisation as an effect of European political integration, a process driven by struggles over the legitimate political and social order that is to prevail in Europe. Firstly, an analytic framework is constructed, drawing on insights from Pierre Bourdieu’s work on similar struggles over nation-stateness. Secondly, the mechanisms identified are used to assess the role played by economic experts and expertise in the process of European political integration. It is argued that concepts arising from economic disciplines, agents educated in economics, and practising economic professionals influence European political integration and have benefited from Europeanisation initiated by this process. Special emphasis is placed on strategies of integrating Europe by law or by market, on governing Europe using economic expertise, on the role played by economic academia in researching and objectifying Europe, and on staffing European institutions with economists.
Is there an ideal time window for language acquisition after which nativelike
representation and processing are unattainable? Although this question has
been heavily debated, no consensus has been reached. Here, we present
evidence for a sensitive period in language development and show that it is
specific to grammar. We conducted a masked priming task with a group of
Turkish-German bilinguals and examined age of acquisition (AoA) effects on
the processing of complex words. We compared a subtle but meaningful
linguistic contrast, that between grammatical inflection and lexical-based
derivation. The results showed a highly selective AoA effect on inflectional
(but not derivational) priming. In addition, the effect displayed a discontinuity
indicative of a sensitive period: Priming from inflected forms was nativelike
when acquisition started before the age of 5 but declined with increasing
AoA. We conclude that the acquisition of morphological rules expressing
morphosyntactic properties is constrained by maturational factors.
Many educational technology proponents support the Technological
Pedagogical Content Knowledge (TPACK) model as a way to
conceptualize teaching with technology, but recent TPACK research
shows a need for empirical studies regarding the development of this
knowledge. This proof-of-concept study applies mixed-methods to
investigate the meta-cognitive awareness produced by teachers who
participate in the Graphic Assessment of TPACK Instrument (GATI).
This process involves creating graphical representations (circles of
differing sizes and the degree of their overlap) that represent what
teachers understand to be their current and aspired TPACK. This study
documented teachers’ explanations during a think-aloud procedure as
they created their GATI figures. The in-depth data from two German
teachers who participated in the process captured the details of their
experience and demonstrated the potential of the GATI to support
teachers in reflecting about their professional knowledge and in
determining their own professional development activities. These
findings will be informative to future pilot studies involving the larger
design of the GATI process, to better understand the role of teachers’
meta-conceptual awareness, and to better ascertain how the GATI
might be used to support professional development on a larger scale.
Between-school variation in students' achievement, motivation, affect, and learning strategies
(2017)
To plan group-randomized trials where treatment conditions are assigned to schools, researchers need design parameters that provide information about between-school differences in outcomes as well as the amount of variance that can be explained by covariates at the student (L1) and school (L2) levels. Most previous research has offered these parameters for U.S. samples and for achievement as the outcome. This paper and the online supplementary materials provide design parameters for 81 countries in three broad outcome categories (achievement, affect and motivation, and learning strategies) for domain-general and domain-specific (mathematics, reading, and science) measures. Sociodemographic characteristics were used as covariates. Data from representative samples of 15-year-old students stemmed from five cycles of the Programme for International Student Assessment (PISA; total number of students/schools: 1,905,147/70,098). Between-school differences as well as the amount of variance explained at L1 and L2 varied widely across countries and educational outcomes, demonstrating the limited generalizability of design parameters across these dimensions. The use of the design parameters to plan group-randomized trials is illustrated.
Coping, taming or solving
(2017)
One of the truisms of policy analysis is that policy problems are
rarely solved. As an ever-increasing number of policy issues are
identified as an inherently ill-structured and intractable type of
wicked problem, the question of what policy analysis sets out
to accomplish has emerged as more central than ever. If solving
wicked problems is beyond reach, research on wicked problems
needs to provide a clearer understanding of the alternatives.
The article identifies and explicates three distinguishable
strategies of problem governance: coping, taming and solving.
It shows that their intellectual premises and practical
implications clearly contrast in core respects. The article argues
that none of the identified strategies of problem governance is
invariably more suitable for dealing with wicked problems.
Rather than advocate for some universally applicable approach
to the governance of wicked problems, the article asks under
what conditions different ways of governing wicked problems
are analytically reasonable and normatively justified. It
concludes that a more systematic assessment of alternative
approaches of problem governance requires a reorientation of
the debate away from the conception of wicked problems as a
singular type toward the more focused analysis of different
dimensions of problem wickedness.
This study investigates the comprehension of wh-questions in individuals with aphasia (IWA) speaking Turkish, a non-wh-movement language, and German, a wh-movement language. We examined six German-speaking and 11 Turkish-speaking IWA using picture-pointing tasks. Findings from our experiments show that the Turkish IWA responded more accurately to both object who and object which questions than to subject questions, while the German IWA performed better for subject which questions than in all other conditions. Using random forest models, a machine learning technique used in tree-structured classification, on the individual data revealed that both the Turkish and German IWA’s response accuracy is largely predicted by the presence of overt and unambiguous case marking. We discuss our results with regard to different theoretical approaches to the comprehension of wh-questions in aphasia.
Nowadays, graph data models are employed, when relationships between entities have to be stored and are in the scope of queries. For each entity, this graph data model locally stores relationships to adjacent entities. Users employ graph queries to query and modify these entities and relationships. These graph queries employ graph patterns to lookup all subgraphs in the graph data that satisfy certain graph structures. These subgraphs are called graph pattern matches. However, this graph pattern matching is NP-complete for subgraph isomorphism. Thus, graph queries can suffer a long response time, when the number of entities and relationships in the graph data or the graph patterns increases.
One possibility to improve the graph query performance is to employ graph views that keep ready graph pattern matches for complex graph queries for later retrieval. However, these graph views must be maintained by means of an incremental graph pattern matching to keep them consistent with the graph data from which they are derived, when the graph data changes. This maintenance adds subgraphs that satisfy a graph pattern to the graph views and removes subgraphs that do not satisfy a graph pattern anymore from the graph views.
Current approaches for incremental graph pattern matching employ Rete networks. Rete networks are discrimination networks that enumerate and maintain all graph pattern matches of certain graph queries by employing a network of condition tests, which implement partial graph patterns that together constitute the overall graph query. Each condition test stores all subgraphs that satisfy the partial graph pattern. Thus, Rete networks suffer high memory consumptions, because they store a large number of partial graph pattern matches. But, especially these partial graph pattern matches enable Rete networks to update the stored graph pattern matches efficiently, because the network maintenance exploits the already stored partial graph pattern matches to find new graph pattern matches. However, other kinds of discrimination networks exist that can perform better in time and space than Rete networks. Currently, these other kinds of networks are not used for incremental graph pattern matching.
This thesis employs generalized discrimination networks for incremental graph pattern matching. These discrimination networks permit a generalized network structure of condition tests to enable users to steer the trade-off between memory consumption and execution time for the incremental graph pattern matching. For that purpose, this thesis contributes a modeling language for the effective definition of generalized discrimination networks. Furthermore, this thesis contributes an efficient and scalable incremental maintenance algorithm, which updates the (partial) graph pattern matches that are stored by each condition test. Moreover, this thesis provides a modeling evaluation, which shows that the proposed modeling language enables the effective modeling of generalized discrimination networks. Furthermore, this thesis provides a performance evaluation, which shows that a) the incremental maintenance algorithm scales, when the graph data becomes large, and b) the generalized discrimination network structures can outperform Rete network structures in time and space at the same time for incremental graph pattern matching.
Data profiling is the computer science discipline of analyzing a given dataset for its metadata. The types of metadata range from basic statistics, such as tuple counts, column aggregations, and value distributions, to much more complex structures, in particular inclusion dependencies (INDs), unique column combinations (UCCs), and functional dependencies (FDs). If present, these statistics and structures serve to efficiently store, query, change, and understand the data. Most datasets, however, do not provide their metadata explicitly so that data scientists need to profile them.
While basic statistics are relatively easy to calculate, more complex structures present difficult, mostly NP-complete discovery tasks; even with good domain knowledge, it is hardly possible to detect them manually. Therefore, various profiling algorithms have been developed to automate the discovery. None of them, however, can process datasets of typical real-world size, because their resource consumptions and/or execution times exceed effective limits.
In this thesis, we propose novel profiling algorithms that automatically discover the three most popular types of complex metadata, namely INDs, UCCs, and FDs, which all describe different kinds of key dependencies. The task is to extract all valid occurrences from a given relational instance. The three algorithms build upon known techniques from related work and complement them with algorithmic paradigms, such as divide & conquer, hybrid search, progressivity, memory sensitivity, parallelization, and additional pruning to greatly improve upon current limitations. Our experiments show that the proposed algorithms are orders of magnitude faster than related work. They are, in particular, now able to process datasets of real-world, i.e., multiple gigabytes size with reasonable memory and time consumption.
Due to the importance of data profiling in practice, industry has built various profiling tools to support data scientists in their quest for metadata. These tools provide good support for basic statistics and they are also able to validate individual dependencies, but they lack real discovery features even though some fundamental discovery techniques are known for more than 15 years. To close this gap, we developed Metanome, an extensible profiling platform that incorporates not only our own algorithms but also many further algorithms from other researchers. With Metanome, we make our research accessible to all data scientists and IT-professionals that are tasked with data profiling. Besides the actual metadata discovery, the platform also offers support for the ranking and visualization of metadata result sets.
Being able to discover the entire set of syntactically valid metadata naturally introduces the subsequent task of extracting only the semantically meaningful parts. This is challenge, because the complete metadata results are surprisingly large (sometimes larger than the datasets itself) and judging their use case dependent semantic relevance is difficult. To show that the completeness of these metadata sets is extremely valuable for their usage, we finally exemplify the efficient processing and effective assessment of functional dependencies for the use case of schema normalization.
Natural products and their derivatives have always been a source of drug leads. In particular, bacterial compounds have played an important role in drug development, for example in the field of antibiotics. A decrease in the discovery of novel leads from natural sources and the hope of finding new leads through the generation of large libraries of drug-like compounds by combinatorial chemistry aimed at specific molecular targets drove the pharmaceutical companies away from research on natural products. However, recent technological advances in genetics, bioinformatics and analytical chemistry have revived the interest in natural products. The ribosomally synthesized and post-translationally modified peptides (RiPPs) are a group of natural products generated by the action of post-translationally modifying enzymes on precursor peptides translated from mRNA by ribosomes. The great substrate promiscuity exhibited by many of the enzymes from RiPP biosynthetic pathways have led to the generation of hundreds of novel synthetic and semisynthetic variants, including variants carrying non-canonical amino acids (ncAAs). The microviridins are a family of RiPPs characterized by their atypical tricyclic structure composed of lactone and lactam rings, and their activity as serine protease inhibitors. The generalities of their biosynthetic pathway have already been described, however, the lack of information on details such as the protease responsible for cleaving off the leader peptide from the cyclic core peptide has impeded the fast and cheap production of novel microviridin variants. In the present work, knowledge on leader peptide activation of enzymes from other RiPP families has been extrapolated to the microviridin family, making it possible to bypass the need of a leader peptide. This feature allowed for the exploitation of the microviridin biosynthetic machinery for the production of novel variants through the establishment of an efficient one-pot in vitro platform. The relevance of this chemoenzymatic approach has been exemplified by the synthesis of novel potent serine protease inhibitors from both rationally-designed peptide libraries and bioinformatically predicted microviridins. Additionally, new structure-activity relationships (SARs) could be inferred by screening microviridin intermediates. The significance of this technique was further demonstrated by the simple incorporation of ncAAs into the microviridin scaffold.
Recognizing, understanding, and responding to quantities are considerable skills for human beings. We can easily communicate quantities, and we are extremely efficient in adapting our behavior to numerical related tasks. One usual task is to compare quantities. We also use symbols like digits in numerical-related tasks. To solve tasks including digits, we must to rely on our previously learned internal number representations.
This thesis elaborates on the process of number comparison with the use of noisy mental representations of numbers, the interaction of number and size representations and how we use mental number representations strategically. For this, three studies were carried out.
In the first study, participants had to decide which of two presented digits was numerically larger. They had to respond with a saccade in the direction of the anticipated answer. Using only a small set of meaningfully interpretable parameters, a variant of random walk models is described that accounts for response time, error rate, and variance of response time for the full matrix of 72 digit pairs. In addition, the used random walk model predicts a numerical distance effect even for error response times and this effect clearly occurs in the observed data. In relation to corresponding correct answers error responses were systematically faster. However, different from standard assumptions often made in random walk models, this account required that the distributions of step sizes of the induced random walks be asymmetric to account for this asymmetry between correct and incorrect responses.
Furthermore, the presented model provides a well-defined framework to investigate the nature and scale (e.g., linear vs. logarithmic) of the mapping of numerical magnitude onto its internal representation. In comparison of the fits of proposed models with linear and logarithmic mapping, the logarithmic mapping is suggested to be prioritized.
Finally, we discuss how our findings can help interpret complex findings (e.g., conflicting speed vs. accuracy trends) in applied studies that use number comparison as a well-established diagnostic tool. Furthermore, a novel oculomotoric effect is reported, namely the saccadic overschoot effect. The participants responded by saccadic eye movements and the amplitude of these saccadic responses decreases with numerical distance.
For the second study, an experimental design was developed that allows us to apply the signal detection theory to a task where participants had to decide whether a presented digit was physically smaller or larger. A remaining question is, whether the benefit in (numerical magnitude – physical size) congruent conditions is related to a better perception than in incongruent conditions. Alternatively, the number-size congruency effect is mediated by response biases due to numbers magnitude. The signal detection theory is a perfect tool to distinguish between these two alternatives. It describes two parameters, namely sensitivity and response bias. Changes in the sensitivity are related to the actual task performance due to real differences in perception processes whereas changes in the response bias simply reflect strategic implications as a stronger preparation (activation) of an anticipated answer. Our results clearly demonstrate that the number-size congruency effect cannot be reduced to mere response bias effects, and that genuine sensitivity gains for congruent number-size pairings contribute to the number-size congruency effect.
Third, participants had to perform a SNARC task – deciding whether a presented digit was odd or even. Local transition probability of irrelevant attributes (magnitude) was varied while local transition probability of relevant attributes (parity) and global probability occurrence of each stimulus were kept constantly. Participants were quite sensitive in recognizing the underlying local transition probability of irrelevant attributes. A gain in performance was observed for actual repetitions of the irrelevant attribute in relation to changes of the irrelevant attribute in high repetition conditions compared to low repetition conditions. One interpretation of these findings is that information about the irrelevant attribute (magnitude) in the previous trial is used as an informative precue, so that participants can prepare early processing stages in the current trial, with the corresponding benefits and costs typical of standard cueing studies.
Finally, the results reported in this thesis are discussed in relation to recent studies in numerical cognition.
Proteins are molecules that are essential for life and carry out an enormous number of functions in organisms. To this end, they change their conformation and bind to other molecules. However, the interplay between conformational change and binding is not fully understood. In this work, this interplay is investigated with molecular dynamics (MD) simulations of the protein-peptide system Mdm2-PMI and by analysis of data from relaxation experiments.
The central task it to uncover the binding mechanism, which is described by the sequence of (partial) binding events and conformational change events including their probabilities. In the simplest case, the binding mechanism is described by a two-step model: binding followed by conformational change or conformational change followed by binding. In the general case, longer sequences with multiple conformational changes and partial binding events are possible as well as parallel pathways that differ in their sequences of events. The theory of Markov state models (MSMs) provides the theoretical framework in which all these cases can be modeled. For this purpose, MSMs are estimated in this work from MD data, and rate equation models, which are related to MSMs, are inferred from experimental relaxation data.
The MD simulation and Markov modeling of the PMI-Mdm2 system shows that PMI and Mdm2 can bind via multiple pathways. A main result of this work is a dissociation rate on the order of one event per second, which was calculated using Markov modeling and is in agreement with experiment. So far, dissociation rates and transition rates of this magnitude have only been calculated with methods that speed up transitions by acting with time-dependent, external forces on the binding partners. The simulation technique developed in this work, in contrast, allows the estimation of dissociation rates from the combination of free energy calculation and direct MD simulation of the fast binding process. Two new statistical estimators TRAM and TRAMMBAR are developed to estimate a MSM from the joint data of both simulation types.
In addition, a new analysis technique for time-series data from chemical relaxation experiments is developed in this work. It allows to identify one of the above-mentioned two-step mechanisms as the mechanism that underlays the data. The new method is valid for a broader range of concentrations than previous methods and therefore allows to choose the concentrations such that the mechanism can be uniquely identified. It is successfully tested with data for the binding of recoverin to a rhodopsin kinase peptide.
Start-up incentives targeted at unemployed individuals have become an important tool of the Active Labor Market Policy (ALMP) to fight unemployment in many countries in recent years. In contrast to traditional ALMP instruments like training measures, wage subsidies, or job creation schemes, which are aimed at reintegrating unemployed individuals into dependent employment, start-up incentives are a fundamentally different approach to ALMP, in that they intend to encourage and help unemployed individuals to exit unemployment by entering self-employment and, thus, by creating their own jobs. In this sense, start-up incentives for unemployed individuals serve not only as employment and social policy to activate job seekers and combat unemployment but also as business policy to promote entrepreneurship. The corresponding empirical literature on this topic so far has been mainly focused on the individual labor market perspective, however. The main part of the thesis at hand examines the new start-up subsidy (“Gründungszuschuss”) in Germany and consists of four empirical analyses that extend the existing evidence on start-up incentives for unemployed individuals from multiple perspectives and in the following directions:
First, it provides the first impact evaluation of the new start-up subsidy in Germany. The results indicate that participation in the new start-up subsidy has significant positive and persistent effects on both reintegration into the labor market as well as the income profiles of participants, in line with previous evidence on comparable German and international programs, which emphasizes the general potential of start-up incentives as part of the broader ALMP toolset. Furthermore, a new innovative sensitivity analysis of the applied propensity score matching approach integrates findings from entrepreneurship and labor market research about the key role of an individual’s personality on start-up decision, business performance, as well as general labor market outcomes, into the impact evaluation of start-up incentives. The sensitivity analysis with regard to the inclusion and exclusion of usually unobserved personality variables reveals that differences in the estimated treatment effects are small in magnitude and mostly insignificant. Consequently, concerns about potential overestimation of treatment effects in previous evaluation studies of similar start-up incentives due to usually unobservable personality variables are less justified, as long as the set of observed control variables is sufficiently informative (Chapter 2).
Second, the thesis expands our knowledge about the longer-term business performance and potential of subsidized businesses arising from the start-up subsidy program. In absolute terms, the analysis shows that a relatively high share of subsidized founders successfully survives in the market with their original businesses in the medium to long run. The subsidy also yields a “double dividend” to a certain extent in terms of additional job creation. Compared to “regular”, i.e., non-subsidized new businesses founded by non-unemployed individuals in the same quarter, however, the economic and growth-related impulses set by participants of the subsidy program are only limited with regard to employment growth, innovation activity, or investment. Further investigations of possible reasons for these differences show that differential business growth paths of subsidized founders in the longer run seem to be mainly limited by higher restrictions to access capital and by unobserved factors, such as less growth-oriented business strategies and intentions, as well as lower (subjective) entrepreneurial persistence. Taken together, the program has only limited potential as a business and entrepreneurship policy intended to induce innovation and economic growth (Chapters 3 and 4).
And third, an empirical analysis on the level of German regional labor markets yields that there is a high regional variation in subsidized start-up activity relative to overall new business formation. The positive correlation between regular start-up intensity and the share among all unemployed individuals who participate in the start-up subsidy program suggests that (nascent) unemployed founders also profit from the beneficial effects of regional entrepreneurship capital. Moreover, the analysis of potential deadweight and displacement effects from an aggregated regional perspective emphasizes that the start-up subsidy for unemployed individuals represents a market intervention into existing markets, which affects incumbents and potentially produces inefficiencies and market distortions. This macro perspective deserves more attention and research in the future (Chapter 5).
The Cauchy problem for the linearised Einstein equation and the Goursat problem for wave equations
(2017)
In this thesis, we study two initial value problems arising in general relativity. The first is the Cauchy problem for the linearised Einstein equation on general globally hyperbolic spacetimes, with smooth and distributional initial data. We extend well-known results by showing that given a solution to the linearised constraint equations of arbitrary real Sobolev regularity, there is a globally defined solution, which is unique up to addition of gauge solutions. Two solutions are considered equivalent if they differ by a gauge solution. Our main result is that the equivalence class of solutions depends continuously on the corre- sponding equivalence class of initial data. We also solve the linearised constraint equations in certain cases and show that there exist arbitrarily irregular (non-gauge) solutions to the linearised Einstein equation on Minkowski spacetime and Kasner spacetime.
In the second part, we study the Goursat problem (the characteristic Cauchy problem) for wave equations. We specify initial data on a smooth compact Cauchy horizon, which is a lightlike hypersurface. This problem has not been studied much, since it is an initial value problem on a non-globally hyperbolic spacetime. Our main result is that given a smooth function on a non-empty, smooth, compact, totally geodesic and non-degenerate Cauchy horizon and a so called admissible linear wave equation, there exists a unique solution that is defined on the globally hyperbolic region and restricts to the given function on the Cauchy horizon. Moreover, the solution depends continuously on the initial data. A linear wave equation is called admissible if the first order part satisfies a certain condition on the Cauchy horizon, for example if it vanishes. Interestingly, both existence of solution and uniqueness are false for general wave equations, as examples show. If we drop the non-degeneracy assumption, examples show that existence of solution fails even for the simplest wave equation. The proof requires precise energy estimates for the wave equation close to the Cauchy horizon. In case the Ricci curvature vanishes on the Cauchy horizon, we show that the energy estimates are strong enough to prove local existence and uniqueness for a class of non-linear wave equations. Our results apply in particular to the Taub-NUT spacetime and the Misner spacetime. It has recently been shown that compact Cauchy horizons in spacetimes satisfying the null energy condition are necessarily smooth and totally geodesic. Our results therefore apply if the spacetime satisfies the null energy condition and the Cauchy horizon is compact and non-degenerate.
Anthropogenically amplified erosion leads to increased fine-grained sediment input into the fluvial system in the 15.000 km2 Kharaa River catchment in northern Mongolia and constitutes a major stressing factor for the aquatic ecosystem. This study uniquely combines the application of intensive monitoring, source fingerprinting and catchment modelling techniques to allow for the comparison of the credibility and accuracy of each single method. High-resolution discharge data were used in combination with daily suspended solid measurements to calculate the suspended sediment budget and compare it with estimations of the sediment budget model SedNet. The comparison of both techniques showed that the development of an overall sediment budget with SedNet was possible, yielding results in the same order of magnitude (20.3 kt a- 1 and 16.2 kt a- 1).
Radionuclide sediment tracing, using Be-7, Cs-137 and Pb-210 was applied to differentiate sediment sources for particles < 10μm from hillslope and riverbank erosion and showed that riverbank erosion generates 74.5% of the suspended sediment load, whereas surface erosion contributes 21.7% and gully erosion only 3.8%. The contribution of the single subcatchments of the Kharaa to the suspended sediment load was assessed based on their variation in geochemical composition (e.g. in Ti, Sn, Mo, Mn, As, Sr, B, U, Ca and Sb). These variations were used for sediment source discrimination with geochemical composite fingerprints based on Genetic Algorithm driven Discriminant Function Analysis, the Kruskal–Wallis H-test and Principal Component Analysis. The contributions of the individual sub-catchment varied from 6.4% to 36.2%, generally showing higher contributions from the sub-catchments in the middle, rather than the upstream portions of the study area.
The results indicate that river bank erosion generated by existing grazing practices of livestock is the main cause for elevated fine sediment input. Actions towards the protection of the headwaters and the stabilization of the river banks within the middle reaches were identified as the highest priority. Deforestation and by lodging and forest fires should be prevented to avoid increased hillslope erosion in the mountainous areas. Mining activities are of minor importance for the overall catchment sediment load but can constitute locally important point sources for particular heavy metals in the fluvial system.
Self-adaptive data quality
(2017)
Carrying out business processes successfully is closely linked to the quality of the data inventory in an organization. Lacks in data quality lead to problems: Incorrect address data prevents (timely) shipments to customers. Erroneous orders lead to returns and thus to unnecessary effort. Wrong pricing forces companies to miss out on revenues or to impair customer satisfaction. If orders or customer records cannot be retrieved, complaint management takes longer. Due to erroneous inventories, too few or too much supplies might be reordered.
A special problem with data quality and the reason for many of the issues mentioned above are duplicates in databases. Duplicates are different representations of same real-world objects in a dataset. However, these representations differ from each other and are for that reason hard to match by a computer. Moreover, the number of required comparisons to find those duplicates grows with the square of the dataset size. To cleanse the data, these duplicates must be detected and removed. Duplicate detection is a very laborious process. To achieve satisfactory results, appropriate software must be created and configured (similarity measures, partitioning keys, thresholds, etc.). Both requires much manual effort and experience.
This thesis addresses automation of parameter selection for duplicate detection and presents several novel approaches that eliminate the need for human experience in parts of the duplicate detection process.
A pre-processing step is introduced that analyzes the datasets in question and classifies their attributes semantically. Not only do these annotations help understanding the respective datasets, but they also facilitate subsequent steps, for example, by selecting appropriate similarity measures or normalizing the data upfront. This approach works without schema information.
Following that, we show a partitioning technique that strongly reduces the number of pair comparisons for the duplicate detection process. The approach automatically finds particularly suitable partitioning keys that simultaneously allow for effective and efficient duplicate retrieval. By means of a user study, we demonstrate that this technique finds partitioning keys that outperform expert suggestions and additionally does not need manual configuration. Furthermore, this approach can be applied independently of the attribute types.
To measure the success of a duplicate detection process and to execute the described partitioning approach, a gold standard is required that provides information about the actual duplicates in a training dataset. This thesis presents a technique that uses existing duplicate detection results and crowdsourcing to create a near gold standard that can be used for the purposes above. Another part of the thesis describes and evaluates strategies how to reduce these crowdsourcing costs and to achieve a consensus with less effort.
Swearing in a public place
(2017)
The paper deals with the usage of swear words on the online forum "reddit". Three research questions are dealt with:
How often are swear words used?
How are these swear words received by other users?
Does the topic of the conversation have an influence on the reception and amount of usage of swear words?
The corpus from which the results are taken comprises almost 900 million words. The words are taken from February 2017. Compared to other, similar studies, the corpus is considerably larger and contempory.
In addition, the theoretical part discusses the linguistic basics of swear words. These include concepts such as the theory of politeness, the topic of taboos and its corresponding words and censorship. This is done to explain the factors that influence the use and application of swear words and to explain why swearwords are so special in comparison to other word groups. In addition, further research results from other corpora are presented and compared with the results afterwards. This includes corpora that are also composed of online communication, as well as corpora that reproduce spoken language. The results from all the corpora presented deal with results from the English language.
The results of this study indicate that the swear words on "reddit" are used approximately as often as they are on other platforms. The perception of these swear words is mostly positive, which suggests that the use of swear words on "reddit" is not perceived as impolite. In addition, an influence of the discussion topic on the frequency and reception of swear words could be determined.
In this work the human AOX1 was characterized and detailed aspects regarding the expression, the enzyme kinetics and the production of reactive oxygen species (ROS) were investigated. The hAOX1 is a cytosolic enzyme belonging to the molybdenum hydroxylase family. Its catalytically active form is a homodimer with a molecular weight of 300 kDa. Each monomer (150 kDa) consists of three domains: a N-terminal domain (20 kDa) containing two [2Fe-2S] clusters, a 40 kDa intermediate domain containing a flavin adenine dinucleotide (FAD), and a C-terminal domain (85 kDa) containing the substrate binding pocket and the molybdenum cofactor (Moco). The hAOX1 has an emerging role in the metabolism and pharmacokinetics of many drugs, especially aldehydes and N- heterocyclic compounds.
In this study, the hAOX1 was hetereogously expressed in E. coli TP1000 cells, using a new codon optimized gene sequence which improved the expressed protein yield of around 10-fold compared to the previous expression systems for this enzyme. To increase the catalytic activity of hAOX1, an in vitro chemical sulfuration was performed to favor the insertion of the equatorial sulfido ligand at the Moco with consequent increased enzymatic activity of around 10-fold. Steady-state kinetics and inhibition studies were performed using several substrates, electron acceptors and inhibitors. The recombinant hAOX1 showed higher catalytic activity when molecular oxygen was used as electron acceptor. The highest turn over values were obtained with phenanthridine as substrate. Inhibition studies using thioridazine (phenothiazine family), in combination with structural studies performed in the group of Prof. M.J. Romão, Nova Universidade de Lisboa, showed a new inhibition site located in proximity of the dimerization site of hAOX1. The inhibition mode of thioridazine resulted in a noncompetitive inhibition type. Further inhibition studies with loxapine, a thioridazine-related molecule, showed the same type of inhibition. Additional inhibition studies using DCPIP and raloxifene were carried out.
Extensive studies on the FAD active site of the hAOX1 were performed. Twenty new hAOX1 variants were produced and characterized. The hAOX1 variants generated in this work were divided in three groups: I) hAOX1 single nucleotide polymorphisms (SNP) variants; II) XOR- FAD loop hAOX1 variants; III) additional single point hAOX1 variants. The hAOX1 SNP variants G46E, G50D, G346R, R433P, A439E, K1231N showed clear alterations in their catalytic activity, indicating a crucial role of these residues into the FAD active site and in relation to the overall reactivity of hAOX1.
Furthermore, residues of the bovine XOR FAD flexible loop (Q423ASRREDDIAK433) were introduced in the hAOX1. FAD loop hAOX1 variants were produced and characterized for their stability and catalytic activity. Especially the variants hAOX1 N436D/A437D/L438I, N436D/A437D/L438I/I440K and Q434R/N436D/A437D/L438I/I440K showed decreased catalytic activity and stability. hAOX1 wild type and variants were tested for reactivity toward NADH but no reaction was observed.
Additionally, the hAOX1 wild type and variants were tested for the generation of reactive oxygen species (ROS). Interestingly, one of the SNP variants, hAOX1 L438V, showed a high ratio of superoxide prodction. This result showed a critical role for the residue Leu438 in the mechanism of oxygen radicals formation by hAOX1. Subsequently, further hAOX1 variants having the mutated Leu438 residue were produced. The variants hAOX1 L438A, L438F and L438K showed superoxide overproduction of around 85%, 65% and 35% of the total reducing equivalent obtained from the substrate oxidation.
The results of this work show for the first time a characterization of the FAD active site of the hAOX1, revealing the importance of specific residues involved in the generation of ROS and effecting the overall enzymatic activity of hAOX1. The hAOX1 SNP variants presented here indicate that those allelic variations in humans might cause alterations ROS balancing and clearance of drugs in humans.
A Case for Serious Play
(2017)
In this thesis, stochastic dynamics modelling collective motions of populations, one of the most mysterious type of biological phenomena, are considered. For a system of N particle-like individuals, two kinds of asymptotic behaviours are studied : ergodicity and flocking properties, in long time, and propagation of chaos, when the number N of agents goes to infinity. Cucker and Smale, deterministic, mean-field kinetic model for a population without a hierarchical structure is the starting point of our journey : the first two chapters are dedicated to the understanding of various stochastic dynamics it inspires, with random noise added in different ways. The third chapter, an attempt to improve those results, is built upon the cluster expansion method, a technique from statistical mechanics. Exponential ergodicity is obtained for a class of non-Markovian process with non-regular drift. In the final part, the focus shifts onto a stochastic system of interacting particles derived from Keller and Segel 2-D parabolicelliptic model for chemotaxis. Existence and weak uniqueness are proven.
The general purpose of this systematic review was to summarize, structure and evaluate the findings on automatic evaluations of exercising. Studies were eligible for inclusion if they reported measuring automatic evaluations of exercising with an implicit measure and assessed some kind of exercise variable. Fourteen nonexperimental and six experimental studies (out of a total N = 1,928) were identified and rated by two independent reviewers. The main study characteristics were extracted and the grade of evidence for each study evaluated. First, results revealed a large heterogeneity in the applied measures to assess automatic evaluations of exercising and the exercise variables. Generally, small to large-sized significant relations between automatic evaluations of exercising and exercise variables were identified in the vast majority of studies. The review offers a systematization of the various examined exercise variables and prompts to differentiate more carefully between actually observed exercise behavior (proximal exercise indicator) and associated physiological or psychological variables (distal exercise indicator). Second, a lack of transparent reported reflections on the differing theoretical basis leading to the use of specific implicit measures was observed. Implicit measures should be applied purposefully, taking into consideration the individual advantages or disadvantages of the measures. Third, 12 studies were rated as providing first-grade evidence (lowest grade of evidence), five represent second-grade and three were rated as third-grade evidence. There is a dramatic lack of experimental studies, which are essential for illustrating the cause-effect relation between automatic evaluations of exercising and exercise and investigating under which conditions automatic evaluations of exercising influence behavior. Conclusions about the necessity of exercise interventions targeted at the alteration of automatic evaluations of exercising should therefore not be drawn too hastily.
The present study explored teachers' perspectives on one specific type of acceleration, namely, grade skipping. In addition, we investigated the extent to which teachers' beliefs about students' academic, motivational, and social development after grade skipping may explain teachers' acceptance of this accelerative strategy. Moreover, we examined whether teachers' acceptance is linked to their decisions about using this intervention. Using data from the PARS project, which included 316 teachers from 18 secondary schools in the German federal state of North Rhine-Westphalia, we assessed teachers' acceptance, beliefs, and perceived knowledge about grade skipping using 4-point rating scales. Teachers also reported whether they had advised a student to skip a grade. Multilevel regression analyses indicated that teachers' beliefs about students' social, motivational, and academic development largely explained their acceptance. Teachers who showed a higher level of acceptance and perceived knowledge were more likely to have recommended grade skipping before. Educational implications are discussed.
Plant functional traits reflect individual and community ecological strategies. They allow the detection of directional changes in community dynamics and ecosystemic processes, being an additional tool to assess biodiversity than species richness. Analysis of functional patterns in plant communities provides mechanistic insight into biodiversity alterations due to anthropogenic activity. Although studies have considered of either anthropogenic management or nutrient availability on functional traits in temperate grasslands, studies combining effects of both drivers are scarce. Here, we assessed the impacts of management intensity (fertilization, mowing, grazing), nutrient stoichiometry (C, N, P, K), and vegetation composition on community-weighted means (CWMs) and functional diversity (Rao's Q) from seven plant traits in 150 grasslands in three regions in Germany, using data of 6 years. Land use and nutrient stoichiometry accounted for larger proportions of model variance of CWM and Rao's Q than species richness and productivity. Grazing affected all analyzed trait groups; fertilization and mowing only impacted generative traits. Grazing was clearly associated with nutrient retention strategies, that is, investing in durable structures and production of fewer, less variable seed. Phenological variability was increased. Fertilization and mowing decreased seed number/mass variability, indicating competition-related effects. Impacts of nutrient stoichiometry on trait syndromes varied. Nutrient limitation (large N:P, C:N ratios) promoted species with conservative strategies, that is, investment in durable plant structures rather than fast growth, fewer seed, and delayed flowering onset. In contrast to seed mass, leaf-economics variability was reduced under P shortage. Species diversity was positively associated with the variability of generative traits. Synthesis. Here, land use, nutrient availability, species richness, and plant functional strategies have been shown to interact complexly, driving community composition, and vegetation responses to management intensity. We suggest that deeper understanding of underlying mechanisms shaping community assembly and biodiversity will require analyzing all these parameters.
The timing and location of the two largest earthquakes of the 21st century (Sumatra, 2004 and Tohoku 2011, events) greatly surprised the scientific community, indicating that the deformation processes that precede and follow great megathrust earthquakes remain enigmatic. During these phases before and after the earthquake a combination of multi-scale complex processes are acting simultaneously: Stresses built up by long-term tectonic motions are modified by sudden jerky deformations during earthquakes, before being restored by multiple ensuing relaxation processes.
This thesis details a cross-scale thermomechanical model developed with the aim of simulating the entire subduction process from earthquake (1 minute) to million years’ time scale, excluding only rupture propagation. The model employs elasticity, non-linear transient viscous rheology, and rate-and-state friction. It generates spontaneous earthquake sequences, and, by using an adaptive time-step algorithm, recreates the deformation process as observed naturally over single and multiple seismic cycles. The model is thoroughly tested by comparing results to those from known high- resolution solutions of generic modeling setups widely used in modeling of rupture propagation. It is demonstrated, that while not modeling rupture propagation explicitly, the modeling procedure correctly recognizes the appearance of instability (earthquake) and correctly simulates the cumulative slip at a fault during great earthquake by means of a quasi-dynamic approximation.
A set of 2D models is used to study the effects of non-linear transient rheology on the postseismic processes following great earthquakes. Our models predict that the viscosity in the mantle wedge drops by 3 to 4 orders of magnitude during a great earthquake with magnitude above 9. This drop in viscosity results in spatial scales and timings of the relaxation processes following the earthquakes that are significantly different to previous estimates. These models replicate centuries long seismic cycles exhibited by the greatest earthquakes (like the Great Chile 1960 Earthquake) and are consistent with the major features of postseismic surface displacements recorded after the Great Tohoku Earthquake.
The 2D models are also applied to study key factors controlling maximum magnitudes of earthquakes in subduction zones. Even though methods of instrumentally observing earthquakes at subduction zones have rapidly improved in recent decades, the characteristic recurrence interval of giant earthquakes (Mw>8.5) is much larger than the currently available observational record and therefore the necessary conditions for giant earthquakes are not clear. Statistical studies have recognized the importance of the slab shape and its surface roughness, state of the strain of the upper plate and thickness of sediments filling the trenches. In this thesis we attempt to explain these observations and to identify key controlling parameters. We test a set of 2D models representing great earthquake seismic cycles at known subduction zones with various known geometries, megathrust friction coefficients, and convergence rates implemented. We found that low-angle subduction (large effect) and thick sediments in the subduction channel (smaller effect) are the fundamental necessary conditions for generating giant earthquakes, while the change of subduction velocity from 10 to 3.5 cm/yr has a lower effect. Modeling results also suggest that having thick sediments in the subduction channel causes low static friction, resulting in neutral or slightly compressive deformation in the overriding plate for low-angle subduction zones. These modeling results agree well with observations for the largest earthquakes. The model predicts the largest possible earthquakes for subduction zones of given dipping angles. The predicted maximum magnitudes exactly threshold magnitudes of all known giant earthquakes of 20th and 21st centuries.
The clear limitation of most of the models developed in the thesis is their 2D nature. Development of 3D models with comparable resolution and complexity will require significant advances in numerical techniques. Nevertheless, we conducted a series of low-resolution 3D models to study the interaction between two large asperities at a subduction interface separated by an aseismic gap of varying width. The novelty of the model is that it considers behavior of the asperities during multiple seismic cycles. As expected, models show that an aseismic gap with a narrow width could not prevent rupture propagation from one asperity to another, and that rupture always crosses the entire model. When the gap becomes too wide, asperities do not interact anymore and rupture independently. However, an interesting mode of interaction was observed in the model with an intermediate width of the aseismic gap: In this model the asperities began to stably rupture in anti-phase following multiple seismic cycles. These 3D modeling results, while insightful, must be considered preliminary because of the limitations in resolution.
The technique developed in this thesis for cross-scale modeling of seismic cycles can be used to study the effects of multiple seismic cycles on the long-term deformation of the upper plate. The technique can be also extended to the case of continental transform faults and for the advanced 3D modeling of specific subduction zones. This will require further development of numerical techniques and adaptation of the existing advanced highly scalable parallel codes like LAMEM and ASPECT.
Magnetotactic bacteria possess an intracellular structure called the magnetosome chain. Magnetosome chains contain nano−particles of iron crystals enclosed by a membrane and aligned on a cytoskeletal filament. Due to the presence of the magnetosome chains, magnetotactic bacteria are able to orient and swim along the magnetic field lines. A detailed study of structural properties of magnetosome chains in magnetotactic bacteria has primary scientific interests. It can provide more insight into the formation of the cytoskeleton in bacteria. In this thesis, we develop a new framework to study the structural properties of magnetosome chains in magnetotactic bacteria.
First, we address the bending stiffness of magnetosome chains resulting from two main contributions: the magnetic interactions of magnetosome particles and the bending stiffness of the cytoskeletal filament to which the magnetosomes are anchored. Our analysis indicates that the linear configuration of magnetosome particles without the stabilisation to the cytoskeleton may close to ring like structures, with no net magnetic moment, which thus can not perform as a compass in cellular navigation. As a result we think that one of the roles of the filament is to stabilize the linear configuration against ring closure.
We then investigate the equilibrium configurations of magnetosome particles including linear chain and closed−ring structures. We notably observe that for the formation of a stable linear structure on the cytoskeletal filament, presence of a binding energy is needed. In the presence of external stimuli the stability of the magnetosome chain is due to the internal dipole−dipole interactions, the stiffness and the binding energy of the protein structure connecting the magnetosome particles to the filament. Our observations, during and after the treatment of the magnetosome chain with the external magnetic field substantiates the stabilisation of magnetosome chains to the cytoskeletal filament by proteinous linkers and the dynamic feature of these structures.
Finally, we employ our model to study the FMR spectra of magnetosome chains in a single cell of magnetotactic bacteria. We explore the effect of magnetocrystalline anisotropy in three-fold symmetry observed in FMR spectra and the peculiarity of different spectra arisen from different mutants of these bacteria.
Detection and Kirchhoff-type migration of seismic events by use of a new characteristic function
(2017)
The classical method of seismic event localization is based on the picking of body wave arrivals, ray tracing and inversion of travel time data. Travel time picks with small uncertainties are required to produce reliable and accurate results with this kind of source localization. Hence recordings, with a low Signal-to-Noise Ratio (SNR) cannot be used in a travel time based inversion. Low SNR can be related with weak signals from distant and/or low magnitude sources as well as with a high level of ambient noise. Diffraction stacking is considered as an alternative seismic event localization method that enables also the processing of low SNR recordings by mean of stacking the amplitudes of seismograms along a travel time function. The location of seismic event and its origin time are determined based on the highest stacked amplitudes (coherency) of the image function. The method promotes an automatic processing since it does not need travel time picks as input data.
However, applying diffraction stacking may require longer computation times if only limited computer resources are used. Furthermore, a simple diffraction stacking of recorded amplitudes could possibly fail to locate the seismic sources if the focal mechanism leads to complex radiation patterns which typically holds for both natural and induced seismicity.
In my PhD project, I have developed a new work flow for the localization of seismic events which is based on a diffraction stacking approach. A parallelized code was implemented for the calculation of travel time tables and for the determination of an image function to reduce computation time. In order to address the effects from complex source radiation patterns, I also suggest to compute diffraction stacking from a characteristic function (CF) instead of stacking the original wave form data. A new CF, which is called in the following mAIC (modified from Akaike Information Criterion) is proposed. I demonstrate that, the performance of the mAIC does not depend on the chosen length of the analyzed time window and that both P- and S-wave onsets can be detected accurately. To avoid cross-talk between P- and S-waves due to inaccurate velocity models, I separate the P- and S-waves from the mAIC function by making use of polarization attributes. Then, eventually the final image function is represented by the largest eigenvalue as a result of the covariance analysis between P- and S-image functions. Before applying diffraction stacking, I also apply seismogram denoising by using Otsu thresholding in the time-frequency domain.
Results from synthetic experiments show that the proposed diffraction stacking provides reliable results even from seismograms with low SNR=1. Tests with different presentations of the synthetic seismograms (displacement, velocity, and acceleration) shown that, acceleration seismograms deliver better results in case of high SNR, whereas displacement seismograms provide more accurate results in case of low SNR recordings. In another test, different measures (maximum amplitude, other statistical parameters) were used to determine the source location in the final image function. I found that the statistical approach is the preferred method particularly for low SNR.
The work flow of my diffraction stacking method was finally applied to local earthquake data from Sumatra, Indonesia. Recordings from a temporary network of 42 stations deployed for 9 months around the Tarutung pull-apart Basin were analyzed. The seismic event locations resulting from the diffraction stacking method align along a segment of the Sumatran Fault. A more complex distribution of seismicity is imaged within and around the Tarutung Basin. Two lineaments striking N-S were found in the middle of the Tarutung Basin which support independent results from structural geology. These features are interpreted as opening fractures due to local extension. A cluster of seismic events repeatedly occurred in short time which might be related to fluid drainage since two hot springs are observed at the surface near to this cluster.
In the present work side-chain polystyrenes were synthesized and characterized, in order to be applied in multilayer OLEDs fabricated by solution process techniques. Manufacture of optoelectronic devices by solution process techniques is meant to decrease significantly fabrication cost and allow large scale production of such devices.
This dissertation focusses in three series, enveloped in two material classes. The two classes differ to each other in the type of charge transport exhibited, either ambipolar transport or electron transport. All materials were applied in all-organic solution processed green Ir-based devices.
In the first part, a series of ambipolar host materials were developed to transport both charge types, holes and electrons, and be applied especially as matrix for green Ir-based emitters. It was possible to increase devices efficacy by modulating the predominant charge transport type. This was achieved by modification of molecules electron transport part with more electron-deficient heterocycles or by extending the delocalization of the LUMO. Efficiencies up to 28.9 cd/A were observed for all-organic solution-process three layer devices.
In the second part, suitability of triarylboranes and tetraphenylsilanes as electron transport materials was studied. High triplet energies were obtained, up to 2.95 eV, by rational combination of both molecular structures. Although the combination of both elements had a low effect in materials electron transport properties, high efficiencies around 24 cd/A were obtained for the series in all-organic solution-processed two layer devices.
In the last part, benzene and pyridine were chosen as the series electron-transport motif. By controlling the relative pyridine content (RPC) solubility into methanol was induced for polystyrenes with bulky side-chains. Materials with RPC ≥ 0.5 could be deposited orthogonally from solution without harming underlying layers. From the best of our knowledge, this is the first time such materials are applied in this architecture showing moderate efficiencies around 10 cd/A in all-organic solution processed OLEDs.
Overall, the outcome of these studies will actively contribute to the current research on materials for all-solution processed OLEDs.
In this work, a sensor system based on thermoresponsive materials is developed by utilizing a modular approach. By synthesizing three different key monomers containing either a carboxyl, alkene or alkyne end group connected with a spacer to the methacrylic polymerizable unit, a flexible copolymerization strategy has been set up with oligo ethylene glycol methacrylates. This allows to tune the lower critical solution temperature (LCST) of the polymers in aqueous media. The molar masses are variable thanks to the excurse taken in polymerization in ionic liquids thus stretching molar masses from 25 to over 1000 kDa. The systems that were shown shown to be effective in aqueous solution could be immobilized on surfaces by copolymerizing photo crosslinkable units. The immobilized systems were formulated to give different layer thicknesses, swelling ratios and mesh sizes depending on the demand of the coupling reaction.
The coupling of detector units or model molecules is approached via reactions of the click chemistry pool, and the reactions are evaluated on their efficiency under those aspects, too. These coupling reactions are followed by surface plasmon resonance spectroscopy (SPR) to judge efficiency. With these tools at hand, Salmonella saccharides could be selectively detected by SPR. Influenza viruses were detected in solution by turbidimetry in solution as well as by a copolymerized solvatochromic dye to track binding via the changes of the polymers’ fluorescence by said binding event. This effect could also be achieved by utilizing the thermoresponsive behavior. Another demonstrator consists of the detection system bound to a quartz surface, thus allowing the virus detection on a solid carrier.
The experiments show the great potential of combining the concepts of thermoresponsive materials and click chemistry to develop technically simple sensors for large biomolecules and viruses.
The classical Navier-Stokes equations of hydrodynamics are usually written in terms of vector analysis. More promising is the formulation of these equations in the language of differential forms of degree one. In this way the study of Navier-Stokes equations includes the analysis of the de Rham complex. In particular, the Hodge theory for the de Rham complex enables one to eliminate the pressure from the equations. The Navier-Stokes equations constitute a parabolic system with a nonlinear term which makes sense only for one-forms. A simpler model of dynamics of incompressible viscous fluid is given by Burgers' equation. This work is aimed at the study of invariant structure of the Navier-Stokes equations which is closely related to the algebraic structure of the de Rham complex at step 1. To this end we introduce Navier-Stokes equations related to any elliptic quasicomplex of first order differential operators. These equations are quite similar to the classical Navier-Stokes equations including generalised velocity and pressure vectors. Elimination of the pressure from the generalised Navier-Stokes equations gives a good motivation for the study of the Neumann problem after Spencer for elliptic quasicomplexes. Such a study is also included in the work.We start this work by discussion of Lamé equations within the context of elliptic quasicomplexes on compact manifolds with boundary. The non-stationary Lamé equations form a hyperbolic system. However, the study of the first mixed problem for them gives a good experience to attack the linearised Navier-Stokes equations. On this base we describe a class of non-linear perturbations of the Navier-Stokes equations, for which the solvability results still hold.
Import and decomposition of dissolved organic carbon in pre-dams of drinking water reservoirs
(2017)
Dissolved organic carbon (DOC) depicts a key component in the aquatic carbon cycle as well as for drinking water production from surface waters. DOC concentrations increased in water bodies of the northern hemisphere in the last decades, posing ecological consequences and water quality problems. Within the pelagic zone of lakes and reservoirs, the DOC pool is greatly affected by biological activity as DOC is simultaneously produced and decomposed. This thesis aimed for a conceptual understanding of organic carbon cycling and DOC quality changes under differing hydrological and trophic conditions. Further, the occurrence of aquatic priming was investigated, which has been proposed as a potential process facilitating the microbial decomposition of stable allochthonous DOC within the pelagic zone.
To study organic carbon cycling under different hydrological conditions, quantitative and qualitative investigations were carried out in three pre-dams of drinking water reservoirs exhibiting a gradient in DOC concentrations and trophic states. All pre-dams were mainly autotrophic in their epilimnia. Discharge and temperature were identified as the key factors regulating net production and respiration in the upper water layers of the pre-dams. Considerable high autochthonous production was observed during the summer season under higher trophic status and base flow conditions. Up to 30% of the total gained organic carbon was produced within the epilimnia. Consequently, this affected the DOC quality within the pre-dams over the year and enhanced characteristics of algae-derived DOC were observed during base flow in summer. Allochthonous derived DOC dominated at high discharges and oligotrophic conditions when production and respiration were low. These results underline that also small impoundments with typically low water residence times are hotspots of carbon cycling, significantly altering water quality in dependence of discharge conditions, temperature and trophic status. Further, it highlights that these factors need to be considered in future water management as increasing temperatures and altered precipitation patterns are predicted in the context of climate change.
Under base flow conditions, heterotrophic bacteria preferentially utilized older DOC components with a conventional radiocarbon age of 195-395 years before present (i.e. before 1950). In contrast, younger carbon components (modern, i.e. produced after 1950) were mineralized following a storm flow event. This highlights that age and recalcitrance of DOC are independent from each other. To assess the ages of the microbially consumed DOC, a simplified method was developed to recover the respired CO2 from heterotrophic bacterioplankton for carbon isotope analyses (13C, 14C). The advantages of the method comprise the operation of replicate incubations at in-situ temperatures using standard laboratory equipment and thus enabling an application in a broad range of conditions.
Aquatic priming was investigated in laboratory experiments during the microbial decomposition of two terrestrial DOC substrates (peat water and soil leachate). Thereby, natural phytoplankton served as a source of labile organic matter and the total DOC pool increased throughout the experiments due to exudation and cell lysis of the growing phytoplankton. A priming effect for both terrestrial DOC substrates was revealed via carbon isotope analysis and mixing models. Thereby, priming was more pronounced for the peat water than for the soil leachate. This indicates that the DOC source and the amount of the added labile organic matter might influence the magnitude of a priming effect. Additional analysis via high-resolution mass spectrometry revealed that oxidized, unsaturated compounds were more strongly decomposed under priming (i.e. in phytoplankton presence). Given the observed increase in DOC concentrations during the experiments, it can be concluded that aquatic priming is not easily detectable via net concentration changes alone and could be considered as a qualitative effect.
The knowledge gained from this thesis contributes to the understanding of aquatic carbon cycling and demonstrated how DOC dynamics in freshwaters vary with hydrological, seasonal and trophic conditions. It further demonstrated that aquatic priming contributes to the microbial transformation of organic carbon and the observed decay of allochthonous DOC during transport in inland waters.
Mathematical models of bacterial growth have been successfully applied to study the relationship between antibiotic drug exposure and the antibacterial effect. Since these models typically lack a representation of cellular processes and cell physiology, the mechanistic integration of drug action is not possible on the cellular level. The cellular mechanisms of drug action, however, are particularly relevant for the prediction, analysis and understanding of interactions between antibiotics. Interactions are also studied experimentally, however, a lacking consent on the experimental protocol hinders direct comparison of results. As a consequence, contradictory classifications as additive, synergistic or antagonistic are reported in literature.
In the present thesis we developed a novel mathematical model for bacterial growth that integrates cell-level processes into the population growth level. The scope of the model is to predict bacterial growth under antimicrobial perturbation by multiple antibiotics in vitro.
To this end, we combined cell-level data from literature with population growth data for Bacillus subtilis, Escherichia coli and Staphylococcus aureus. The cell-level data described growth-determining characteristics of a reference cell, including the ribosomal concentration and efficiency. The population growth data comprised extensive time-kill curves for clinically relevant antibiotics (tetracycline, chloramphenicol, vancomycin, meropenem, linezolid, including dual combinations).
The new cell-level approach allowed for the first time to simultaneously describe single and combined effects of the aforementioned antibiotics for different experimental protocols, in particular different growth phases (lag and exponential phase). Consideration of ribosomal dynamics and persisting sub-populations explained the decreased potency of linezolid on cultures in the lag phase compared to exponential phase cultures. The model captured growth rate dependent killing and auto-inhibition of meropenem and - also for vancomycin exposure - regrowth of the bacterial cultures due to adaptive resistance development. Stochastic interaction surface analysis demonstrated the pronounced antagonism between meropenem and linezolid to be robust against variation in the growth phase and pharmacodynamic endpoint definition, but sensitive to a change in the experimental duration.
Furthermore, the developed approach included a detailed representation of the bacterial cell-cycle. We used this representation to describe septation dynamics during the transition of a bacterial culture from the exponential to stationary growth phase. Resulting from a new mechanistic understanding of transition processes, we explained the lag time between the increase in cell number and bacterial biomass during the transition from the lag to exponential growth phase. Furthermore, our model reproduces the increased intracellular RNA mass fraction during long term exposure of bacteria to chloramphenicol.
In summary, we contribute a new approach to disentangle the impact of drug effects, assay readout and experimental protocol on antibiotic interactions. In the absence of a consensus on the corresponding experimental protocols, this disentanglement is key to translate information between heterogeneous experiments and also ultimately to the clinical setting.
The reaction of pharmacological active protic ionic liquid tris-(2-hydroxyethyl)ammonium 4-chlorophenylsulfanylacetate H + N(CH 2 CH 2 OH) 3 ∙ ( - OOCCH 2 SC 6 H 4 Cl-4) (1) with zinc or nickel chloride in a ratio of 2:1 affords stable at room temperature powder-like adducts [H + N(CH 2 CH 2 OH) 3 ] 2 ∙ [M(OOCCH 2 SC 6 H 4 Cl-4) 2 Cl 2 ] 2- , M = Zn (2), Ni (3). By recrystallization from aqueous alcohol compound 2 unexpectedly gives Zn(OOCCH 2 SC 6 H 4 Cl-4) 2 ∙ 2H 2 O (4). Unlike 2, compound 3 gives crystals [N(CH 2 CH 2 OH) 3 ] 2 Ni 2+ · [ - OOCCH 2 SC 6 H 4 Cl-4] 2 (5), which have a structure of metallated ionic liquid. The structure of 5 has been proved by X-ray diffraction analysis. It is the first example of the conversion of a protic ionic liquid into potentially biological active metallated ionic liquid (1 → 3 → 5).
Background: Functional abdominal pain (FAP) is not only a highly prevalent disease but also poses a considerable burden on children and their families. Untreated, FAP is highly persistent until adulthood, also leading to an increased risk of psychiatric disorders. Intervention studies underscore the efficacy of cognitive behavioral treatment approaches but are limited in terms of sample size, long-term follow-up data, controls and inclusion of psychosocial outcome data.
Methods/Design: In a multicenter randomized controlled trial, 112 children aged 7 to 12 years who fulfill the Rome III criteria for FAP will be allocated to an established cognitive behavioral training program for children with FAP (n = 56) or to an active control group (focusing on age-appropriate information delivery; n = 56). Randomization occurs centrally, blockwise and is stratified by center. This study is performed in five pediatric gastroenterology outpatient departments. Observer-blind assessments of outcome variables take place four times: pre-, post-, 3- and 12-months post-treatment. Primary outcome is the course of pain intensity and frequency. Secondary endpoints are health-related quality of life, pain-related coping and cognitions, as well as selfefficacy.
Discussion: This confirmatory randomized controlled clinical trial evaluates the efficacy of a cognitive behavioral intervention for children with FAP. By applying an active control group, time and attention processes can be controlled, and long-term follow-up data over the course of one year can be explored.
All life-sustaining processes are ultimately driven by thousands of biochemical reactions occurring in the cells: the metabolism. These reactions form an intricate network which produces all required chemical compounds, i.e., metabolites, from a set of input molecules. Cells regulate the activity through metabolic reactions in a context-specific way; only reactions that are required in a cellular context, e.g., cell type, developmental stage or environmental condition, are usually active, while the rest remain inactive. The context-specificity of metabolism can be captured by several kinds of experimental data, such as by gene and protein expression or metabolite profiles. In addition, these context-specific data can be assimilated into computational models of metabolism, which then provide context-specific metabolic predictions.
This thesis is composed of three individual studies focussing on context-specific experimental data integration into computational models of metabolism. The first study presents an optimization-based method to obtain context-specific metabolic predictions, and offers the advantage of being fully automated, i.e., free of user defined parameters. The second study explores the effects of alternative optimal solutions arising during the generation of context-specific metabolic predictions. These alternative optimal solutions are metabolic model predictions that represent equally well the integrated data, but that can markedly differ. This study proposes algorithms to analyze the space of alternative solutions, as well as some ways to cope with their impact in the predictions.
Finally, the third study investigates the metabolic specialization of the guard cells of the plant Arabidopsis thaliana, and compares it with that of a different cell type, the mesophyll cells. To this end, the computational methods developed in this thesis are applied to obtain metabolic predictions specific to guard cell and mesophyll cells. These cell-specific predictions are then compared to explore the differences in metabolic activity between the two cell types. In addition, the effects of alternative optima are taken into consideration when comparing the two cell types. The computational results indicate a major reorganization of the primary metabolism in guard cells. These results are supported by an independent 13C labelling experiment.
Background: Inferring regulatory interactions between genes from transcriptomics time-resolved data, yielding reverse engineered gene regulatory networks, is of paramount importance to systems biology and bioinformatics studies. Accurate methods to address this problem can ultimately provide a deeper insight into the complexity, behavior, and functions of the underlying biological systems. However, the large number of interacting genes coupled with short and often noisy time-resolved read-outs of the system renders the reverse engineering a challenging task. Therefore, the development and assessment of methods which are computationally efficient, robust against noise, applicable to short time series data, and preferably capable of reconstructing the directionality of the regulatory interactions remains a pressing research problem with valuable applications.
Results: Here we perform the largest systematic analysis of a set of similarity measures and scoring schemes within the scope of the relevance network approach which are commonly used for gene regulatory network reconstruction from time series data. In addition, we define and analyze several novel measures and schemes which are particularly suitable for short transcriptomics time series. We also compare the considered 21 measures and 6 scoring schemes according to their ability to correctly reconstruct such networks from short time series data by calculating summary statistics based on the corresponding specificity and sensitivity. Our results demonstrate that rank and symbol based measures have the highest performance in inferring regulatory interactions. In addition, the proposed scoring scheme by asymmetric weighting has shown to be valuable in reducing the number of false positive interactions. On the other hand, Granger causality as well as information-theoretic measures, frequently used in inference of regulatory networks, show low performance on the short time series analyzed in this study.
Conclusions: Our study is intended to serve as a guide for choosing a particular combination of similarity measures and scoring schemes suitable for reconstruction of gene regulatory networks from short time series data. We show that further improvement of algorithms for reverse engineering can be obtained if one considers measures that are rooted in the study of symbolic dynamics or ranks, in contrast to the application of common similarity measures which do not consider the temporal character of the employed data. Moreover, we establish that the asymmetric weighting scoring scheme together with symbol based measures (for low noise level) and rank based measures (for high noise level) are the most suitable choices.
Background
The kidneys are essential for the metabolism of vitamin A (retinol) and its transport proteins retinol-binding protein 4 (RBP4) and transthyretin. Little is known about changes in serum concentration after living donor kidney transplantation (LDKT) as a consequence of unilateral nephrectomy; although an association of these parameters with the risk of cardiovascular diseases and insulin resistance has been suggested. Therefore we analyzed the concentration of retinol, RBP4, apoRBP4 and transthyretin in serum of 20 living-kidney donors and respective recipients at baseline as well as 6 weeks and 6 months after LDKT.
Results
As a consequence of LDKT, the kidney function of recipients was improved while the kidney function of donors was moderately reduced within 6 weeks after LDKT. With regard to vitamin A metabolism, the recipients revealed higher levels of retinol, RBP4, transthyretin and apoRBP4 before LDKT in comparison to donors. After LDKT, the levels of all four parameters decreased in serum of the recipients, while retinol, RBP4 as well as apoRBP4 serum levels of donors increased and remained increased during the follow-up period of 6 months.
Conclusion
LDKT is generally regarded as beneficial for allograft recipients and not particularly detrimental for the donors. However, it could be demonstrated in this study that a moderate reduction of kidney function by unilateral nephrectomy, resulted in an imbalance of components of vitamin A metabolism with a significant increase of retinol and RBP4 and apoRBP4 concentration in serum of donors.
Background: There is an increasing awareness of the impact of parental risk perception on the weight course of the child and the parent's readiness to engage in preventive efforts, but only less is known about factors related to the parental perception of the right time for the implementation of preventive activities. The aim of this study was to examine parental perceptions of the appropriate time to engage in child weight management strategies, and the factors associated with different weight points at which mothers recognize the need for preventive actions.
Methods: 352 mothers with children aged 2-10 years took part in the study. We assessed mothers' perceptions of the actual and preferred weight status of their child, their ability to identify overweight and knowledge of its associated health risks, as well as perceptions of the right time for action to prevent overweight in their child. A regression analysis was conducted to examine whether demographic and weight related factors as well as the maternal general risk perception were associated with recognizing the need to implement prevention strategies.
Results: Although most of the parents considered a BMI in the 75th to 90th percentile a valid reason to engage in the prevention of overweight, 19% of the mothers were not willing to engage in prevention until their child reached the 97th percentile. Whereas the child's sex and the identification of an elevated BMI were significant predictors for parents' recognition of the 75th percentile as right point to engage in prevention efforts, an inability to recognize physical health risks associated with overweight silhouettes emerged as a significant factor predicting which parents would delay prevention efforts until a child's BMI reached the 97th percentile.
Conclusion: Parental misperceptions of overweight and associated health risks constitute unfavorable conditions for preventive actions. Feedback on the health risks associated with overweight could help increase maternal readiness for change.
Background: Female sperm storage has evolved independently multiple times among vertebrates to control reproduction in response to the environment. In internally fertilising amphibians, female salamanders store sperm in cloacal spermathecae, whereas among anurans sperm storage in oviducts is known only in tailed frogs. Facilitated through extensive field sampling following historical observations we tested for sperm storing structures in the female urogenital tract of fossorial, tropical caecilian amphibians.
Findings: In the oviparous Ichthyophis cf. kohtaoensis, aggregated sperm were present in a distinct region of the posterior oviduct but not in the cloaca in six out of seven vitellogenic females prior to oviposition. Spermatozoa were found most abundantly between the mucosal folds. In relation to the reproductive status decreased amounts of sperm were present in gravid females compared to pre-ovulatory females. Sperm were absent in females past oviposition.
Conclusions: Our findings indicate short-term oviductal sperm storage in the oviparous Ichthyophis cf. kohtaoensis. We assume that in female caecilians exhibiting high levels of parental investment sperm storage has evolved in order to optimally coordinate reproductive events and to increase fitness.
Background: Plasma concentration of retinol is an accepted indicator to assess the vitamin A (retinol) status in cattle. However, the determination of vitamin A requires a time consuming multi-step procedure, which needs specific equipment to perform extraction, centrifugation or saponification prior to high-performance liquid chromatography (HPLC).
Methods: The concentrations of retinol in whole blood (n = 10), plasma (n = 132) and serum (n = 61) were measured by a new rapid cow-side test (iCheck™ FLUORO) and compared with those by HPLC in two independent laboratories in Germany (DE) and Japan (JP).
Results: Retinol concentrations in plasma ranged from 0.033 to 0.532 mg/L, and in serum from 0.043 to 0.360 mg/L (HPLC method). No significant differences in retinol levels were observed between the new rapid cow-side test and HPLC performed in different laboratories (HPLC vs. iCheck™ FLUORO: 0.320 ± 0.047 mg/L vs. 0.333 ± 0.044 mg/L, and 0.240 ± 0.096 mg/L vs. 0.241 ± 0.069 mg/L, lab DE and lab JP, respectively). A similar comparability was observed when whole blood was used (HPLC vs. iCheck™ FLUORO: 0.353 ± 0.084 mg/L vs. 0.341 ± 0.064 mg/L). Results showed a good agreement between both methods based on correlation coefficients of r2 = 0.87 (P < 0.001) and Bland-Altman blots revealed no significant bias for all comparison.
Conclusions: With the new rapid cow-side test (iCheck™ FLUORO) retinol concentrations in cattle can be reliably assessed within a few minutes and directly in the barn using even whole blood without the necessity of prior centrifugation. The ease of the application of the new rapid cow-side test and its portability can improve the diagnostic of vitamin A status and will help to control vitamin A supplementation in specific vitamin A feeding regimes such as used to optimize health status in calves or meat marbling in Japanese Black cattle.
Background: Obesity is not only a highly prevalent disease but also poses a considerable burden on children and their families. Evidence is increasing that a lack of self-regulation skills may play a role in the etiology and maintenance of obesity. Our goal with this currently ongoing trial is to examine whether training that focuses on the enhancement of self-regulation skills may increase the sustainability of a complex lifestyle intervention.
Methods/Design: In a multicenter, prospective, parallel group, randomized controlled superiority trial, 226 obese children and adolescents aged 8 to 16 years will be allocated either to a newly developed computer-training program to improve their self-regulation abilities or to a placebo control group. Randomization occurs centrally and blockwise at a 1:1 allocation ratio for each center. This study is performed in pediatric inpatient rehabilitation facilities specialized in the treatment of obesity. Observer-blind assessments of outcome variables take place at four times: at the beginning of the rehabilitation (pre), at the end of the training in the rehabilitation (post), and 6 and 12 months post-rehabilitation intervention. The primary outcome is the course of BMI-SDS over 1 year after the end of the inpatient rehabilitation. Secondary endpoints are the self-regulation skills. In addition, health-related quality of life, and snack intake will be analyzed.
Discussion: The computer-based training programs might be a feasible and attractive tool to increase the sustainability of the weight loss reached during inpatient rehabilitation.
The interdisciplinary workshop STOCHASTIC PROCESSES WITH APPLICATIONS IN THE NATURAL SCIENCES was held in Bogotá, at Universidad de los Andes from December 5 to December 9, 2016. It brought together researchers from Colombia, Germany, France, Italy, Ukraine, who communicated recent progress in the mathematical research related to stochastic processes with application in biophysics.
The present volume collects three of the four courses held at this meeting by Angelo Valleriani, Sylvie Rœlly and Alexei Kulik.
A particular aim of this collection is to inspire young scientists in setting up research goals within the wide scope of fields represented in this volume.
Angelo Valleriani, PhD in high energy physics, is group leader of the team "Stochastic processes in complex and biological systems" from the Max-Planck-Institute of Colloids and Interfaces, Potsdam.
Sylvie Rœlly, Docteur en Mathématiques, is the head of the chair of Probability at the University of Potsdam.
Alexei Kulik, Doctor of Sciences, is a Leading researcher at the Institute of Mathematics of Ukrainian National Academy of Sciences.
The title compound, erioflorin, C19H24O6 [systematic name: (1aR,3S,4Z,5aR,8aR,9R,10aR)-1a, 2,3,5a, 7,8,8a, 9,10,10a-decahydro-3-hydroxy-4,10a-dimethyl-8-methylidene-7-oxooxireno[5,6] cyclodeca[1,2-b]furan-9-yl methacrylate], is a tricyclic germacrane sesquiterpene lactone, which was isolated from Podanthus mitiqui (L.). The compound crystallizes in the space group P2(1)2(1)2(1), and its molecular structure consists of a methacrylic ester of a ten-membered ring sesquiterpenoid annelated with an epoxide and a butyrolactone. The structure is stabilized by one intramolecular C-H center dot center dot center dot O hydrogen bond. An O-H center dot center dot center dot O hydrogen bond and further C-H center dot center dot center dot O interactions can be observed in the packing.
Background: Although the benefits for health of physical activity (PA) are well documented, the majority of the population is unable to implement present recommendations into daily routine. Mobile health (mHealth) apps could help increase the level of PA. However, this is contingent on the interest of potential users.
Objective: The aim of this study was the explorative, nuanced determination of the interest in mHealth apps with respect to PA among students and staff of a university.
Methods: We conducted a Web-based survey from June to July 2015 in which students and employees from the University of Potsdam were asked about their activity level, interest in mHealth fitness apps, chronic diseases, and sociodemographic parameters.
Results: A total of 1217 students (67.30%, 819/1217; female; 26.0 years [SD 4.9]) and 485 employees (67.5%, 327/485; female; 42.7 years [SD 11.7]) participated in the survey. The recommendation for PA (3 times per week) was not met by 70.1% (340/485) of employees and 52.67% (641/1217) of students. Within these groups, 53.2% (341/641 students) and 44.2% (150/340 employees)—independent of age, sex, body mass index (BMI), and level of education or professional qualification—indicated an interest in mHealth fitness apps.
Conclusions: Even in a younger, highly educated population, the majority of respondents reported an insufficient level of PA. About half of them indicated their interest in training support. This suggests that the use of personalized mobile fitness apps may become increasingly significant for a positive change of lifestyle.
The existence of diverse and active microbial ecosystems in the deep subsurface – a biosphere that was originally considered devoid of life – was discovered in multiple microbiological studies. However, most of the studies are restricted to marine ecosystems, while our knowledge about the microbial communities in the deep subsurface of lake systems and their potentials to adapt to changing environmental conditions is still fragmentary. This doctoral thesis aims to build up a unique data basis for providing the first detailed high-throughput characterization of the deep biosphere of lacustrine sediments and to emphasize how important it is to differentiate between the living and the dead microbial community in deep biosphere studies.
In this thesis, up to 3.6 Ma old sediments (up to 317 m deep) of the El’gygytgyn Crater Lake were examined, which represents the oldest terrestrial climate record of the Arctic. Combining next generation sequencing with detailed geochemical characteristics and other environmental parameters, the microbial community composition was analyzed in regard to changing climatic conditions within the last 3.6 Ma to 1.0 Ma (Pliocene and Pleistocene). DNA from all investigated sediments was successfully extracted and a surprisingly diverse (6,910 OTUs) and abundant microbial community in the El’gygytgyn deep sediments were revealed. The bacterial abundance (10³-10⁶ 16S rRNA copies g⁻¹ sediment) was up to two orders of magnitudes higher than the archaeal abundance (10¹-10⁵) and fluctuates with the Pleistocene glacial/interglacial cyclicality. Interestingly, a strong increase in the microbial diversity with depth was observed (approximately 2.5 times higher diversity in Pliocene sediments compared to Pleistocene sediments). The increase in diversity with depth in the Lake El’gygytgyn is most probably caused by higher sedimentary temperatures towards the deep sediment layers as well as an enhanced temperature-induced intra-lake bioproductivity and higher input of allochthonous organic-rich material during Pliocene climatic conditions. Moreover, the microbial richness parameters follow the general trends of the paleoclimatic parameters, such as the paleo-temperature and paleo-precipitation. The most abundant bacterial representatives in the El’gygytgyn deep biosphere are affiliated with the phyla Proteobacteria, Actinobacteria, Bacteroidetes, and Acidobacteria, which are also commonly distributed in the surrounding permafrost habitats. The predominated taxon was the halotolerant genus Halomonas (in average 60% of the total reads per sample).
Additionally, this doctoral thesis focuses on the live/dead differentiation of microbes in cultures and environmental samples. While established methods (e.g., fluorescence in situ hybridization, RNA analyses) are not applicable to the challenging El’gygytgyn sediments, two newer methods were adapted to distinguish between DNA from live cells and free (extracellular, dead) DNA: the propidium monoazide (PMA) treatment and the cell separation adapted for low amounts of DNA. The applicability of the DNA-intercalating dye PMA was successfully evaluated to mask free DNA of different cultures of methanogenic archaea, which play a major role in the global carbon cycle. Moreover, an optimal procedure to simultaneously treat bacteria and archaea was developed using 130 µM PMA and 5 min of photo-activation with blue LED light, which is also applicable on sandy environmental samples with a particle load of ≤ 200 mg mL⁻¹. It was demonstrated that the soil texture has a strong influence on the PMA treatment in particle-rich samples and that in particular silt and clay-rich samples (e.g., El’gygytgyn sediments) lead to an insufficient shielding of free DNA by PMA. Therefore, a cell separation protocol was used to distinguish between DNA from live cells (intracellular DNA) and extracellular DNA in the El’gygytgyn sediments. While comparing these two DNA pools with a total DNA pool extracted with a commercial kit, significant differences in the microbial composition of all three pools (mean distance of relative abundance: 24.1%, mean distance of OTUs: 84.0%) was discovered. In particular, the total DNA pool covers significantly fewer taxa than the cell-separated DNA pools and only inadequately represents the living community. Moreover, individual redundancy analyses revealed that the microbial community of the intra- and extracellular DNA pool are driven by different environmental factors. The living community is mainly influenced by life-dependent parameters (e.g., sedimentary matrix, water availability), while the extracellular DNA is dependent on the biogenic silica content. The different community-shaping parameters and the fact, that a redundancy analysis of the total DNA pool explains significantly less variance of the microbial community, indicate that the total DNA represents a mixture of signals of the live and dead microbial community.
This work provides the first fundamental data basis of the diversity and distribution of microbial deep biosphere communities of a lake system over several million years. Moreover, it demonstrates the substantial importance of extracellular DNA in old sediments. These findings may strongly influence future environmental community analyses, where applications of live/dead differentiation avoid incorrect interpretations due to a failed extraction of the living microbial community or an overestimation of the past community diversity in the course of total DNA extraction approaches.
Carbohydrate-protein interactions are ubiquitous in nature. They provide the initial molecular contacts in many cell-cell processes as for example immune responses, signal transduction, egg fertilization and infection processes of pathogenic viruses and bacteria. Furthermore, bacteria themselves are infected by bacteriophages, viruses which can cause the bacterial lysis, but do not affect other hosts. The infection process of a bacteriophage involves the specific detection and binding of the bacterium, which can be based on a carbohydrate-protein interaction. The mechanism of specific detection of pathogenic bacteria can thereby be useful for the development of bacteria sensors in the food industry or for tools in diagnostics.
Bacteriophages of the Podoviridae family use tailspike proteins for the specific detection of enteritis causing bacteria as Escherichia coli, Salmonella spp. or Shigella flexneri. The tailspike protein provides the first contact by binding to the carbohydrate containing O-antigen part of lipopolysaccharide in the Gram-negative cell wall. After binding to O-antigen repeating units, the enzymatic activity of tailspike proteins leads to cleavage of the carbohydrate chains, which enables the bacteriophage to approach the bacterial surface for DNA injection. Tailspike proteins thereby exhibit a relatively low affinity to the oligosaccharide structures of O-antigen due to the necessary binding, cleavage and release cycle, compared for example to antibodies. In this work it was aimed to study the determinants that influence carbohydrate affinity in the extended TSP binding grooves. This is a prerequisite to design a high-affinity tailspike protein based bacteria sensor.
For this purpose the tailspike protein of the bacteriophage Sf6 (Sf6 TSP) was used, which specifically binds Shigella flexneri Y O-antigen with two tetrasaccharide repeating units at the intersubunits of the trimeric β-helix protein. The Sf6 TSP endorhamnosidase cleaves the O-antigen, which leads to an octasaccharide as the main product. The binding affinity of inactive Sf6 TSP towards polysaccharide was characterized by fluorescence titration experiments and surface plasmon resonance (SPR).
Moreover, cysteine mutations were introduced into the Sf6 TSP binding site for the covalent thiol-coupling of an environment-sensitive fluorescent label to obtain a sensor for Shigella flexneri Y based on TSP-O-antigen recognition. This sensor showed a more than 100 % amplitude increase of a visible light fluorescence upon the binding of a polysaccharide test solution. Improvements of the TSP sensor can be achieved by increasing the tailspike affinity towards the O-antigen. Therefore molecular dynamics simulations evaluating ligand flexibility, hydrogen bond occupancies and water network distributions were used for affinity prediction on the available cysteine mutants of Sf6 TSP. The binding affinities were experimentally analyzed by SPR. This combined computational and experimental set-up for the design of a high-affinity carbohydrate binding protein could successfully distinguish strongly increased and decreased affinities of single amino acid mutants.
A thermodynamically and structurally well characterized set of another tailspike protein HK620 TSP with high-affinity mutants was used to evaluate the influence of water molecules on binding affinity. The free enthalpy of HK620 TSP oligosaccharide complex formation thereby either derived from the replacement of a conserved water molecule or by immobilization of two water molecules upon ligand binding. Furthermore, the enthalpic and entropic contributions of water molecules in a hydrophobic binding pocket could be assigned by free energy calculations. The findings in this work can be helpful for the improvement of carbohydrate docking and carbohydrate binding protein engineering algorithms in the future.
We analyze an inverse noisy regression model under random design with the aim of estimating the unknown target function based on a given set of data, drawn according to some unknown probability distribution. Our estimators are all constructed by kernel methods, which depend on a Reproducing Kernel Hilbert Space structure using spectral regularization methods.
A first main result establishes upper and lower bounds for the rate of convergence under a given source condition assumption, restricting the class of admissible distributions. But since kernel methods scale poorly when massive datasets are involved, we study one example for saving computation time and memory requirements in more detail. We show that Parallelizing spectral algorithms also leads to minimax optimal rates of convergence provided the number of machines is chosen appropriately.
We emphasize that so far all estimators depend on the assumed a-priori smoothness of the target function and on the eigenvalue decay of the kernel covariance operator, which are in general unknown. To obtain good purely data driven estimators constitutes the problem of adaptivity which we handle for the single machine problem via a version of the Lepskii principle.
One of the most striking features of recent public sector reform in Europe is privatization. This development raises questions of accountability: By whom and for what are managers of private for-profit organizations delivering public goods held accountable? Analyzing accountability mechanisms through the lens of an institutional organizational approach and on the empirical basis of hospital privatization in Germany, the article contributes to the empirical and theoretical understanding of public accountability of private actors. The analysis suggests that accountability is not declining but rather multiplying. The shifts in the locus and content of accountability cause organizational stress for private hospitals.
The predictions of two contrasting approaches to the acquisition of transitive relative clauses were tested within the same groups of German-speaking participants aged from 3 to 5 years old. The input frequency approach predicts that object relative clauses with inanimate heads (e.g., the pullover that the man is scratching) are comprehended earlier and more accurately than those with an animate head (e.g., the man that the boy is scratching). In contrast, the structural intervention approach predicts that object relative clauses with two full NP arguments mismatching in number (e.g., the man that the boys are scratching) are comprehended earlier and more accurately than those with number-matching NPs (e.g., the man that the boy is scratching). These approaches were tested in two steps. First, we ran a corpus analysis to ensure that object relative clauses with number-mismatching NPs are not more frequent than object relative clauses with number-matching NPs in child directed speech. Next, the comprehension of these structures was tested experimentally in 3-, 4-, and 5-year-olds respectively by means of a color naming task. By comparing the predictions of the two approaches within the same participant groups, we were able to uncover that the effects predicted by the input frequency and by the structural intervention approaches co-exist and that they both influence the performance of children on transitive relative clauses, but in a manner that is modulated by age. These results reveal a sensitivity to animacy mismatch already being demonstrated by 3-year-olds and show that animacy is initially deployed more reliably than number to interpret relative clauses correctly. In all age groups, the animacy mismatch appears to explain the performance of children, thus, showing that the comprehension of frequent object relative clauses is enhanced compared to the other conditions. Starting with 4-year-olds but especially in 5-year-olds, the number mismatch supported comprehension—a facilitation that is unlikely to be driven by input frequency. Once children fine-tune their sensitivity to verb agreement information around the age of four, they are also able to deploy number marking to overcome the intervention effects. This study highlights the importance of testing experimentally contrasting theoretical approaches in order to characterize the multifaceted, developmental nature of language acquisition.