Refine
Has Fulltext
- yes (343) (remove)
Year of publication
- 2022 (343) (remove)
Document Type
- Postprint (158)
- Doctoral Thesis (137)
- Working Paper (20)
- Monograph/Edited Volume (11)
- Article (9)
- Bachelor Thesis (2)
- Conference Proceeding (2)
- Part of Periodical (2)
- Habilitation Thesis (1)
- Master's Thesis (1)
Language
- English (343) (remove)
Keywords
- machine learning (7)
- climate change (6)
- COVID-19 (5)
- Klimawandel (5)
- exercise (4)
- obesity (4)
- Innovation (3)
- Modellierung (3)
- Röntgenspektroskopie (3)
- bibliometric analysis (3)
Institute
- Institut für Biochemie und Biologie (49)
- Extern (48)
- Strukturbereich Kognitionswissenschaften (33)
- Hasso-Plattner-Institut für Digital Engineering GmbH (30)
- Institut für Physik und Astronomie (30)
- Institut für Geowissenschaften (27)
- Fachgruppe Volkswirtschaftslehre (24)
- Center for Economic Policy Analysis (CEPA) (20)
- Institut für Chemie (18)
- Department Sport- und Gesundheitswissenschaften (14)
In its practical outlook, the interdisciplinary-driven colonial discourse theory is often criticized for its totalizing tendencies regarding the structure of the exam-ined discourse and the power relations prevailing in this framework. As a result of this structural totalization, the concerned subjects got disempowered and de-graded to mere passive objects incapable of raising their voices within the dis-course. Based on this justified criticism, this thesis investigates the role colonial subjects played in the emergence, the distribution, as well as in questioning and critiquing of the colonial discourse during the initial phase of British colonialism in West Africa. The focal point lies on three themes relevant to the period be-tween 1874 and 1914: The Ashanti-Wars, the creation of an educational system, and the issue of the so-called "Europeanized Africans." Newspapers published by the colonial elite serve as the central source material in order to reconstruct Afri-can perspectives on these subjects. First, the discursive trajectory of the first two themes will be reconstructed and then shown why the initial support of the elite gradually declined towards the end of the century. Eventually, the analyzed tendencies culminated in the emergence of the "African Regeneration" discourse, which was able to reverse the colonial discourse's basic assumptions, at least on a theoretical level. Consequently, the Africans were displayed as the "civilizer" of Europe. On the structural level, however, this discourse likewise employed a to-talizing picture of African and European societies, respectively.
Data stream processing systems (DSPSs) are a key enabler to integrate continuously generated data, such as sensor measurements, into enterprise applications. DSPSs allow to steadily analyze information from data streams, e.g., to monitor manufacturing processes and enable fast reactions to anomalous behavior. Moreover, DSPSs continuously filter, sample, and aggregate incoming streams of data, which reduces the data size, and thus data storage costs.
The growing volumes of generated data have increased the demand for high-performance DSPSs, leading to a higher interest in these systems and to the development of new DSPSs. While having more DSPSs is favorable for users as it allows choosing the system that satisfies their requirements the most, it also introduces the challenge of identifying the most suitable DSPS regarding current needs as well as future demands. Having a solution to this challenge is important because replacements of DSPSs require the costly re-writing of applications if no abstraction layer is used for application development. However, quantifying performance differences between DSPSs is a difficult task. Existing benchmarks fail to integrate all core functionalities of DSPSs and lack tool support, which hinders objective result comparisons. Moreover, no current benchmark covers the combination of streaming data with existing structured business data, which is particularly relevant for companies.
This thesis proposes a performance benchmark for enterprise stream processing called ESPBench. With enterprise stream processing, we refer to the combination of streaming and structured business data. Our benchmark design represents real-world scenarios and allows for an objective result comparison as well as scaling of data. The defined benchmark query set covers all core functionalities of DSPSs. The benchmark toolkit automates the entire benchmark process and provides important features, such as query result validation and a configurable data ingestion rate.
To validate ESPBench and to ease the use of the benchmark, we propose an example implementation of the ESPBench queries leveraging the Apache Beam software development kit (SDK). The Apache Beam SDK is an abstraction layer designed for developing stream processing applications that is applied in academia as well as enterprise contexts. It allows to run the defined applications on any of the supported DSPSs. The performance impact of Apache Beam is studied in this dissertation as well. The results show that there is a significant influence that differs among DSPSs and stream processing applications. For validating ESPBench, we use the example implementation of the ESPBench queries developed using the Apache Beam SDK. We benchmark the implemented queries executed on three modern DSPSs: Apache Flink, Apache Spark Streaming, and Hazelcast Jet. The results of the study prove the functioning of ESPBench and its toolkit. ESPBench is capable of quantifying performance characteristics of DSPSs and of unveiling differences among systems.
The benchmark proposed in this thesis covers all requirements to be applied in enterprise stream processing settings, and thus represents an improvement over the current state-of-the-art.
The aim of this dissertation was to conduct a larger-scale cross-linguistic empirical investigation of similarity-based interference effects in sentence comprehension.
Interference studies can offer valuable insights into the mechanisms that are involved in long-distance dependency completion.
Many studies have investigated similarity-based interference effects, showing that syntactic and semantic information are employed during long-distance dependency formation (e.g., Arnett & Wagers, 2017; Cunnings & Sturt, 2018; Van Dyke, 2007, Van Dyke & Lewis, 2003; Van Dyke & McElree, 2011). Nevertheless, there are some important open questions in the interference literature that are critical to our understanding of the constraints involved in dependency resolution.
The first research question concerns the relative timing of syntactic and semantic interference in online sentence comprehension. Only few interference studies have investigated this question, and, to date, there is not enough data to draw conclusions with regard to their time course (Van Dyke, 2007; Van Dyke & McElree, 2011).
Our first cross-linguistic study explores the relative timing of syntactic and semantic interference in two eye-tracking reading experiments that implement the study design used in Van Dyke (2007). The first experiment tests English sentences. The second, larger-sample experiment investigates the two interference types in German.
Overall, the data suggest that syntactic and semantic interference can arise simultaneously during retrieval.
The second research question concerns a special case of semantic interference: We investigate whether cue-based retrieval interference can be caused by semantically similar items which are not embedded in a syntactic structure.
This second interference study builds on a landmark study by Van Dyke & McElree (2006). The study design used in their study is unique in that it is able to pin down the source of interference as a consequence of cue overload during retrieval, when semantic retrieval cues do not uniquely match the retrieval target. Unlike most other interference studies, this design is able to rule out encoding interference as an alternative explanation. Encoding accounts postulate that it is not cue overload at the retrieval site but the erroneous encoding of similar linguistic items in memory that leads to interference (Lewandowsky et al., 2008; Oberauer & Kliegl, 2006). While Van Dyke & McElree (2006) reported cue-based retrieval interference from sentence-external distractors, the evidence for this effect was weak. A subsequent study did not show interference of this type (Van Dyke et al., 2014). Given these inconclusive findings, further research is necessary to investigate semantic cue-based retrieval interference.
The second study in this dissertation provides a larger-scale cross-linguistic investigation of cue-based retrieval interference from sentence-external items. Three larger-sample eye-tracking studies in English, German, and Russian tested cue-based interference in the online processing of filler-gap dependencies. This study further extends the previous research by investigating interference in each language under varying task demands (Logačev & Vasishth, 2016; Swets et al., 2008).
Overall, we see some very modest support for proactive cue-based retrieval interference in English. Unexpectedly, this was observed only under a low task demand. In German and Russian, there is some evidence against the interference effect. It is possible that interference is attenuated in languages with richer case marking.
In sum, the cross-linguistic experiments on the time course of syntactic and semantic interference from sentence-internal distractors support existing evidence of syntactic and semantic interference during sentence comprehension. Our data further show that both types of interference effects can arise simultaneously. Our cross-linguistic experiments investigating semantic cue-based retrieval interference from sentence-external distractors suggest that this type of interference may arise only in specific linguistic contexts.
What are the consequences of unemployment and precarious employment for individuals' health in Europe? What are the moderating factors that may offset (or increase) the health consequences of labor-market risks? How do the effects of these risks vary across different contexts, which differ in their institutional and cultural settings? Does gender, regarded as a social structure, play a role, and how? To answer these questions is the aim of my cumulative thesis. This study aims to advance our knowledge about the health consequences that unemployment and precariousness cause over the life course. In particular, I investigate how several moderating factors, such as gender, the family, and the broader cultural and institutional context, may offset or increase the impact of employment instability and insecurity on individual health.
In my first paper, 'The buffering role of the family in the relationship between job loss and self-perceived health: Longitudinal results from Europe, 2004-2011', I and my co-authors measure the causal effect of job loss on health and the role of the family and welfare states (regimes) as moderating factors. Using EU-SILC longitudinal data (2004-2011), we estimate the probability of experiencing 'bad health' following a transition to unemployment by applying linear probability models and undertake separate analyses for men and women. Firstly, we measure whether changes in the independent variable 'job loss' lead to changes in the dependent variable 'self-rated health' for men and women separately. Then, by adding into the model different interaction terms, we measure the moderating effect of the family, both in terms of emotional and economic support, and how much it varies across different welfare regimes. As an identification strategy, we first implement static fixed-effect panel models, which control for time-varying observables and indirect health selection—i.e., constant unobserved heterogeneity. Secondly, to control for reverse causality and path dependency, we implement dynamic fixed-effect panel models, adding a lagged dependent variable to the model.
We explore the role of the family by focusing on close ties within households: we consider the presence of a stable partner and his/her working status as a source of social and economic support. According to previous literature, having a partner should reduce the stress from adverse events, thanks to the symbolic and emotional dimensions that such a relationship entails, regardless of any economic benefits. Our results, however, suggest that benefits linked to the presence of a (female) partner also come from the financial stability that (s)he can provide in terms of a second income. Furthermore, we find partners' employment to be at least as important as the mere presence of the partner in reducing the negative effect of job loss on the individual's health by maintaining the household's standard of living and decreasing economic strain on the family. Our results are in line with previous research, which has highlighted that some people cope better than others with adverse life circumstances, and the support provided by the family is a crucial resource in that regard.
We also reported an important interaction between the family and the welfare state in moderating the health consequences of unemployment, showing how the compensation effect of the family varies across welfare regimes. The family plays a decisive role in cushioning the adverse consequences of labor market risks in Southern and Eastern welfare states, characterized by less developed social protection systems and –especially the Southern – high level of familialism.
The first paper also found important gender differences concerning job loss, family and welfare effects. Of particular interest is the evidence suggesting that health selection works differently for men and women, playing a more prominent role for women than for men in explaining the relationship between job loss and self-perceived health. The second paper, 'Gender roles and selection mechanisms across contexts: A comparative analysis of the relationship between unemployment, self-perceived health, and gender.' investigates more in-depth the gender differential in health driven by unemployment.
Being a highly contested issue in literature, we aim to study whether men are more penalized than women or the other way around and the mechanisms that may explain the gender difference. To do that, we rely on two theoretical arguments: the availability of alternative roles and social selection. The first argument builds on the idea that men and women may compensate for the detrimental health consequences of unemployment through the commitment to 'alternative roles,' which can provide for the resources needed to fulfill people's socially constructed needs. Notably, the availability of alternative options depends on the different positions that men and women have in society.
Further, we merge the availability of the 'alternative roles' argument with the health selection argument. We assume that health selection could be contingent on people's social position as defined by gender and, thus, explain the gender differential in the relationship between unemployment and health. Ill people might be less reluctant to fall or remain (i.e., self-select) in unemployment if they have alternative roles. In Western societies, women generally have more alternative roles than men and thus more discretion in their labor market attachment. Therefore, health selection should be stronger for them, explaining why unemployment is less menace for women than for their male counterparts.
Finally, relying on the idea of different gender regimes, we extended these arguments to comparison across contexts. For example, in contexts where being a caregiver is assumed to be women's traditional and primary roles and the primary breadwinner role is reserved to men, unemployment is less stigmatized, and taking up alternative roles is more socially accepted for women than for men (Hp.1). Accordingly, social (self)selection should be stronger for women than for men in traditional contexts, where, in the case of ill-health, the separation from work is eased by the availability of alternative roles (Hp.2).
By focusing on contexts that are representative of different gender regimes, we implement a multiple-step comparative approach. Firstly, by using EU-SILC longitudinal data (2004-2015), our analysis tests gender roles and selection mechanisms for Sweden and Italy, representing radically different gender regimes, thus providing institutional and cultural variation. Then, we limit institutional heterogeneity by focusing on Germany and comparing East- and West-Germany and older and younger cohorts—for West-Germany (SOEP data 1995-2017). Next, to assess the differential impact of unemployment for men and women, we compared (unemployed and employed) men with (unemployed and employed) women. To do so, we calculate predicted probabilities and average marginal effect from two distinct random-effects probit models. Our first step is estimating random-effects models that assess the association between unemployment and self-perceived health, controlling for observable characteristics. In the second step, our fully adjusted model controls for both direct and indirect selection. We do this using dynamic correlated random-effects (CRE) models. Further, based on the fully adjusted model, we test our hypotheses on alternative roles (Hp.1) by comparing several contexts – models are estimated separately for each context. For this hypothesis, we pool men and women and include an interaction term between unemployment and gender, which has the advantage to allow for directly testing whether gender differences in the effect of unemployment exist and are statistically significant. Finally, we test the role of selection mechanisms (Hp.2), using the KHB method to compare coefficients across nested nonlinear models. Specifically, we test the role of selection for the relationship between unemployment and health by comparing the partially-adjusted and fully-adjusted models. To allow selection mechanisms to operate differently between genders, we estimate separate models for men and women.
We found support to our first hypotheses—the context where people are embedded structures the relationship between unemployment, health, and gender. We found no gendered effect of unemployment on health in the egalitarian context of Sweden. Conversely, in the traditional context of Italy, we observed substantive and statistically significant gender differences in the effect of unemployment on bad health, with women suffering less than men. We found the same pattern for comparing East and West Germany and younger and older cohorts in West Germany.
On the contrary, our results did not support our theoretical argument on social selection. We found that in Sweden, women are more selected out of employment than men. In contrast, in Italy, health selection does not seem to be the primary mechanism behind the gender differential—Italian men and women seem to be selected out of employment to the same extent. Namely, we do not find any evidence that health selection is stronger for women in more traditional countries (Hp2), despite the fact that the institutional and the cultural context would offer them a more comprehensive range of 'alternative roles' relative to men. Moreover, our second hypothesis is also rejected in the second and third comparisons, where the cross-country heterogeneity is reduced to maximize cultural differences within the same institutional context. Further research that addresses selection into inactivity is needed to evaluate the interplay between selection and social roles across gender regimes.
While the health consequences of unemployment have been on the research agenda for a pretty long time, the interest in precarious employment—defined as the linking of the vulnerable worker to work that is characterized by uncertainty and insecurity concerning pay, the stability of the work arrangement, limited access to social benefits, and statutory protections—has emerged only later. Since the 80s, scholars from different disciplines have raised concerns about the social consequences of de-standardization of employment relationships. However, while work has become undoubtedly more precarious, very little is known about its causal effect on individual health and the role of gender as a moderator. These questions are at the core of my third paper : 'Bad job, bad health? A longitudinal analysis of the interaction between precariousness, gender and self-perceived health in Germany'. Herein, I investigate the multidimensional nature of precarious employment and its causal effect on health, particularly focusing on gender differences.
With this paper, I aim at overcoming three major shortcomings of earlier studies: The first one regards the cross-sectional nature of data that prevents the authors from ruling out unobserved heterogeneity as a mechanism for the association between precarious employment and health. Indeed, several unmeasured individual characteristics—such as cognitive abilities—may confound the relationship between precarious work and health, leading to biased results. Secondly, only a few studies have directly addressed the role of gender in shaping the relationship. Moreover, available results on the gender differential are mixed and inconsistent: some found precarious employment being more detrimental for women's health, while others found no gender differences or stronger negative association for men. Finally, previous attempts to an empirical translation of the employment precariousness (EP) concept have not always been coherent with their theoretical framework. EP is usually assumed to be a multidimensional and continuous phenomenon; it is characterized by different dimensions of insecurity that may overlap in the same job and lead to different "degrees of precariousness." However, researchers have predominantly focused on one-dimensional indicators—e.g., temporary employment, subjective job insecurity—to measure EP and study the association with health. Besides the fact that this approach partially grasps the phenomenon's complexity, the major problem is the inconsistency of evidence that it has produced. Indeed, this line of inquiry generally reveals an ambiguous picture, with some studies finding substantial adverse effects of temporary over permanent employment, while others report only minor differences.
To measure the (causal) effect of precarious work on self-rated health and its variation by gender, I focus on Germany and use four waves from SOEP data (2003, 2007, 2011, and 2015). Germany is a suitable context for my study. Indeed, since the 1980s, the labor market and welfare system have been restructured in many ways to increase the German economy's competitiveness in the global market. As a result, the (standard) employment relationship has been de-standardized: non-standard and atypical employment arrangements—i.e., part-time work, fixed-term contracts, mini-jobs, and work agencies—have increased over time while wages have lowered, even among workers with standard work. In addition, the power of unions has also fallen over the last three decades, leaving a large share of workers without collective protection. Because of this process of de-standardization, the link between wage employment and strong social rights has eroded, making workers more powerless and more vulnerable to labor market risks than in the past. EP refers to this uneven distribution of power in the employment relationship, which can be detrimental to workers' health. Indeed, by affecting individuals' access to power and other resources, EP puts precarious workers at risk of experiencing health shocks and influences their ability to gain and accumulate health advantages (Hp.1).
Further, the focus on Germany allows me to investigate my second research question on the gender differential. Germany is usually regarded as a traditionalist gender regime: a context characterized by a configuration of roles. Here, being a caregiver is assumed to be women's primary role, whereas the primary breadwinner role is reserved for men. Although many signs of progress have been made over the last decades towards a greater equalization of opportunities and more egalitarianism, the breadwinner model has barely changed towards a modified version. Thus, women usually take on the double role of workers (the so-called secondary earner) and caregivers, and men still devote most of their time to paid work activities. Moreover, the overall upward trend towards more egalitarian gender ideologies has leveled off over the last decades, moving notably towards more traditional gender ideologies.
In this setting, two alternative hypotheses are possible. Firstly, I assume that the negative relationship between EP and health is stronger for women than for men. This is because women are systematically more disadvantaged than men in the public and private spheres of life, having less access to formal and informal sources of power. These gender-related power asymmetries may interact with EP-related power asymmetries resulting in a stronger effect of EP on women's health than on men's health (Hp.2).
An alternative way of looking at the gender differential is to consider the interaction that precariousness might have with men's and women's gender identities. According to this view, the negative relationship between EP and health is weaker for women than for men (Hp.2a). In a society with a gendered division of labor and a strong link between masculine identities and stable and well-rewarded job—i.e., a job that confers the role of primary family provider—a male worker with precarious employment might violate the traditional male gender role. Men in precarious jobs may perceive themselves (and by others) as possessing a socially undesirable characteristic, which conflicts with the stereotypical idea of themselves as the male breadwinner. Engaging in behaviors that contradict stereotypical gender identity may decrease self-esteem and foster feelings of inferiority, helplessness, and jealousy, leading to poor health.
I develop a new indicator of EP that empirically translates a definition of EP as a multidimensional and continuous phenomenon. I assume that EP is a latent construct composed of seven dimensions of insecurity chosen according to the theory and previous empirical research: Income insecurity, social insecurity, legal insecurity, employment insecurity, working-time insecurity, representation insecurity, worker's vulnerability. The seven dimensions are proxied by eight indicators available in the four waves of the SOEP dataset. The EP composite indicator is obtained by performing a multiple correspondence analysis (MCA) on the eight indicators. This approach aims to construct a summary scale in which all dimensions contribute jointly to the measured experience of precariousness and its health impact.
Further, the relationship between EP and 'general self-perceived health' is estimated by applying ordered probit random-effects estimators and calculating average marginal effect (further AME). Then, to control for unobserved heterogeneity, I implement correlated random-effects models that add to the model the within-individual means of the time-varying independent variables. To test the significance of the gender differential, I add an interaction term between EP and gender in the fully adjusted model in the pooled sample.
My correlated random-effects models showed EP's negative and substantial 'effect' on self-perceived health for both men and women. Although nonsignificant, the evidence seems in line with previous cross-sectional literature. It supports the hypothesis that employment precariousness could be detrimental to workers' health. Further, my results showed the crucial role of unobserved heterogeneity in shaping the health consequences of precarious employment. This is particularly important as evidence accumulates, yet it is still mostly descriptive.
Moreover, my results revealed a substantial difference among men and women in the relationship between EP and health: when EP increases, the risk of experiencing poor health increases much more for men than for women. This evidence falsifies previous theory according to whom the gender differential is contingent on the structurally disadvantaged position of women in western societies. In contrast, they seem to confirm the idea that men in precarious work could experience role conflict to a larger extent than women, as their self-standard is supposed to be the stereotypical breadwinner worker with a good and well-rewarded job. Finally, results from the multiple correspondence analysis contribute to the methodological debate on precariousness, showing that a multidimensional and continuous indicator can express a latent variable of EP.
All in all, complementarities are revealed in the results of unemployment and employment precariousness, which have two implications: Policy-makers need to be aware that the total costs of unemployment and precariousness go far beyond the economic and material realm penetrating other fundamental life domains such as individual health. Moreover, they need to balance the trade-off between protecting adequately unemployed people and fostering high-quality employment in reaction to the highlighted market pressures. In this sense, the further development of a (universalistic) welfare state certainly helps mitigate the adverse health effects of unemployment and, therefore, the future costs of both individuals' health and welfare spending. In addition, the presence of a working partner is crucial for reducing the health consequences of employment instability. Therefore, policies aiming to increase female labor market participation should be promoted, especially in those contexts where the welfare state is less developed.
Moreover, my results support the significance of taking account of a gender perspective in health research. The findings of the three articles show that job loss, unemployment, and precarious employment, in general, have adverse effects on men's health but less or absent consequences for women's health. Indeed, this suggests the importance of labor and health policies that consider and further distinguish the specific needs of the male and female labor force in Europe. Nevertheless, a further implication emerges: the health consequences of employment instability and de-standardization need to be investigated in light of the gender arrangements and the transforming gender relationships in specific cultural and institutional contexts. My results indeed seem to suggest that women's health advantage may be a transitory phenomenon, contingent on the predominant gendered institutional and cultural context. As the structural difference between men's and women's position in society is eroded, egalitarianism becomes the dominant normative status, so will probably be the gender difference in the health consequences of job loss and precariousness. Therefore, while gender equality in opportunities and roles is a desirable aspect for contemporary societies and a political goal that cannot be postponed further, this thesis raises a further and maybe more crucial question: What kind of equality should be pursued to provide men and women with both good life quality and equal chances in the public and private spheres? In this sense, I believe that social and labor policies aiming to reduce gender inequality in society should focus on improving women's integration into the labor market, implementing policies targeting men, and facilitating their involvement in the private sphere of life. Equal redistribution of social roles could activate a crucial transformation of gender roles and the cultural models that sustain and still legitimate gender inequality in Western societies.
We use the prolonged Greek crisis as a case study to understand how a lasting economic shock affects the innovation strategies of firms in economies with moderate innovation activities. Adopting the 3-stage CDM model, we explore the link between R&D, innovation, and productivity for different size groups of Greek manufacturing firms during the prolonged crisis. At the first stage, we find that the continuation of the crisis is harmful for the R&D engagement of smaller firms while it increased the willingness for R&D activities among the larger ones. At the second stage, among smaller firms the knowledge production remains unaffected by R&D investments, while among larger firms the R&D decision is positively correlated with the probability of producing innovation, albeit the relationship is weakened as the crisis continues. At the third stage, innovation output benefits only larger firms in terms of labor productivity, while the innovation-productivity nexus is insignificant for smaller firms during the lasting crisis.
Older adults with amnestic mild cognitive impairment (aMCI) who in addition to their memory deficits also suffer from frontal-executive dysfunctions have a higher risk of developing dementia later in their lives than older adults with aMCI without executive deficits and older adults with non-amnestic MCI (naMCI). Handgrip strength (HGS) is also correlated with the risk of cognitive decline in the elderly. Hence, the current study aimed to investigate the associations between HGS and executive functioning in individuals with aMCI, naMCI and healthy controls. Older, right-handed adults with amnestic MCI (aMCI), non-amnestic MCI (naMCI), and healthy controls (HC) conducted a handgrip strength measurement via a handheld dynamometer. Executive functions were assessed with the Trail Making Test (TMT A&B). Normalized handgrip strength (nHGS, normalized to Body Mass Index (BMI)) was calculated and its associations with executive functions (operationalized through z-scores of TMT B/A ratio) were investigated through partial correlation analyses (i.e., accounting for age, sex, and severity of depressive symptoms). A positive and low-to-moderate correlation between right nHGS (rp (22) = 0.364; p = 0.063) and left nHGS (rp (22) = 0.420; p = 0.037) and executive functioning in older adults with aMCI but not in naMCI or HC was observed. Our results suggest that higher levels of nHGS are linked to better executive functioning in aMCI but not naMCI and HC. This relationship is perhaps driven by alterations in the integrity of the hippocampal-prefrontal network occurring in older adults with aMCI. Further research is needed to provide empirical evidence for this assumption.
An important goal in biotechnology and (bio-) medical research is the isolation of single cells from a heterogeneous cell population. These specialised cells are of great interest for bioproduction, diagnostics, drug development, (cancer) therapy and research. To tackle emerging questions, an ever finer differentiation between target cells and non-target cells is required. This precise differentiation is a challenge for a growing number of available methods.
Since the physiological properties of the cells are closely linked to their morphology, it is beneficial to include their appearance in the sorting decision. For established methods, this represents a non addressable parameter, requiring new methods for the identification and isolation of target cells. Consequently, a variety of new flow-based methods have been developed and presented in recent years utilising 2D imaging data to identify target cells within a sample. As these methods aim for high throughput, the devices developed typically require highly complex fluid handling techniques, making them expensive while offering limited image quality.
In this work, a new continuous flow system for image-based cell sorting was developed that uses dielectrophoresis to precisely handle cells in a microchannel. Dielectrophoretic forces are exerted by inhomogeneous alternating electric fields on polarisable particles (here: cells). In the present system, the electric fields can be switched on and off precisely and quickly by a signal generator. In addition to the resulting simple and effective cell handling, the system is characterised by the outstanding quality of the image data generated and its compatibility with standard microscopes. These aspects result in low complexity, making it both affordable and user-friendly.
With the developed cell sorting system, cells could be sorted reliably and efficiently according to their cytosolic staining as well as morphological properties at different optical magnifications. The achieved purity of the target cell population was up to 95% and about 85% of the sorted cells could be recovered from the system. Good agreement was achieved between the results obtained and theoretical considerations. The achieved throughput of the system was up to 12,000 cells per hour. Cell viability studies indicated a high biocompatibility of the system.
The results presented demonstrate the potential of image-based cell sorting using dielectrophoresis. The outstanding image quality and highly precise yet gentle handling of the cells set the system apart from other technologies. This results in enormous potential for processing valuable and sensitive cell samples.
A multidimensional and analytical perspective on Open Educational Practices in the 21st century
(2022)
Participatory approaches to teaching and learning are experiencing a new lease on life in the 21st century as a result of the rapid technology development. Knowledge, practices, and tools can be shared across spatial and temporal boundaries in higher education by means of Open Educational Resources, Massive Open Online Courses, and open-source technologies. In this context, the Open Education Movement calls for new didactic approaches that encourage greater learner participation in formal higher education. Based on a representative literature review and focus group research, in this study an analytical framework was developed that enables researchers and practitioners to assess the form of participation in formal, collaborative teaching and learning practices. The analytical framework is focused on the micro-level of higher education, in particular on the interaction between students and lecturers when organizing the curriculum. For this purpose, the research reflects anew on the concept of participation, taking into account existing stage models for participation in the educational context. These are then brought together with the dimensions of teaching and learning processes, such as methods, objectives and content, etc. This paper aims to make a valuable contribution to the opening up of learning and teaching, and expands the discourse around possibilities for interpreting Open Educational Practices.
Quantifying the extremeness of heavy precipitation allows for the comparison of events. Conventional quantitative indices, however, typically neglect the spatial extent or the duration, while both are important to understand potential impacts. In 2014, the weather extremity index (WEI) was suggested to quantify the extremeness of an event and to identify the spatial and temporal scale at which the event was most extreme. However, the WEI does not account for the fact that one event can be extreme at various spatial and temporal scales. To better understand and detect the compound nature of precipitation events, we suggest complementing the original WEI with a “cross-scale weather extremity index” (xWEI), which integrates extremeness over relevant scales instead of determining its maximum.
Based on a set of 101 extreme precipitation events in Germany, we outline and demonstrate the computation of both WEI and xWEI. We find that the choice of the index can lead to considerable differences in the assessment of past events but that the most extreme events are ranked consistently, independently of the index. Even then, the xWEI can reveal cross-scale properties which would otherwise remain hidden. This also applies to the disastrous event from July 2021, which clearly outranks all other analyzed events with regard to both WEI and xWEI.
While demonstrating the added value of xWEI, we also identify various methodological challenges along the required computational workflow: these include the parameter estimation for the extreme value distributions, the definition of maximum spatial extent and temporal duration, and the weighting of extremeness at different scales. These challenges, however, also represent opportunities to adjust the retrieval of WEI and xWEI to specific user requirements and application scenarios.
A task-based parallel elliptic solver for numerical relativity with discontinuous Galerkin methods
(2022)
Elliptic partial differential equations are ubiquitous in physics. In numerical relativity---the study of computational solutions to the Einstein field equations of general relativity---elliptic equations govern the initial data that seed every simulation of merging black holes and neutron stars. In the quest to produce detailed numerical simulations of these most cataclysmic astrophysical events in our Universe, numerical relativists resort to the vast computing power offered by current and future supercomputers. To leverage these computational resources, numerical codes for the time evolution of general-relativistic initial value problems are being developed with a renewed focus on parallelization and computational efficiency. Their capability to solve elliptic problems for accurate initial data must keep pace with the increasing detail of the simulations, but elliptic problems are traditionally hard to parallelize effectively.
In this thesis, I develop new numerical methods to solve elliptic partial differential equations on computing clusters, with a focus on initial data for orbiting black holes and neutron stars. I develop a discontinuous Galerkin scheme for a wide range of elliptic equations, and a stack of task-based parallel algorithms for their iterative solution. The resulting multigrid-Schwarz preconditioned Newton-Krylov elliptic solver proves capable of parallelizing over 200 million degrees of freedom to at least a few thousand cores, and already solves initial data for a black hole binary about ten times faster than the numerical relativity code SpEC. I also demonstrate the applicability of the new elliptic solver across physical disciplines, simulating the thermal noise in thin mirror coatings of interferometric gravitational-wave detectors to unprecedented accuracy. The elliptic solver is implemented in the new open-source SpECTRE numerical relativity code, and set up to support simulations of astrophysical scenarios for the emerging era of gravitational-wave and multimessenger astronomy.
Different lake systems might reflect different climate elements of climate changes, while the responses of lake systems are also divers, and are not completely understood so far. Therefore, a comparison of lakes in different climate zones, during the high-amplitude and abrupt climate fluctuations of the Last Glacial to Holocene transition provides an exceptional opportunity to investigate distinct natural lake system responses to different abrupt climate changes. The aim of this doctoral thesis was to reconstruct climatic and environmental fluctuations down to (sub-) annual resolution from two different lake systems during the Last Glacial-Interglacial transition (~17 and 11 ka). Lake Gościąż, situated in the temperate central Poland, developed in the Allerød after recession of the Last Glacial ice sheets. The Dead Sea is located in the Levant (eastern Mediterranean) within a steep gradient from sub-humid to hyper-arid climate, and formed in the mid-Miocene. Despite their differences in sedimentation processes, both lakes form annual laminations (varves), which are crucial for studies of abrupt climate fluctuations. This doctoral thesis was carried out within the DFG project PALEX-II (Paleohydrology and Extreme Floods from the Dead Sea ICDP Core) that investigates extreme hydro-meteorological events in the ICDP core in relation to climate changes, and ICLEA (Virtual Institute of Integrated Climate and Landscape Evolution Analyses) that intends to better the understanding of climate dynamics and landscape evolutions in north-central Europe since the Last Glacial. Further, it contributes to the Helmholtz Climate Initiative REKLIM (Regional Climate Change and Humans) Research Theme 3 “Extreme events across temporal and spatial scales” that investigates extreme events using climate data, paleo-records and model-based simulations. The three main aims were to (1) establish robust chronologies of the lakes, (2) investigate how major and abrupt climate changes affect the lake systems, and (3) to compare the responses of the two varved lakes to these hemispheric-scale climate changes.
Robust chronologies are a prerequisite for high-resolved climate and environmental reconstructions, as well as for archive comparisons. Thus, addressing the first aim, the novel chronology of Lake Gościąż was established by microscopic varve counting and Bayesian age-depth modelling in Bacon for a non-varved section, and was corroborated by independent age constrains from 137Cs activity concentration measurements, AMS radiocarbon dating and pollen analysis. The varve chronology reaches from the late Allerød until AD 2015, revealing more Holocene varves than a previous study of Lake Gościąż suggested. Varve formation throughout the complete Younger Dryas (YD) even allowed the identification of annually- to decadal-resolved leads and lags in proxy responses at the YD transitions.
The lateglacial chronology of the Dead Sea (DS) was thus far mainly based on radiocarbon and U/Th-dating. In the unique ICDP core from the deep lake centre, continuous search for cryptotephra has been carried out in lateglacial sediments between two prominent gypsum deposits – the Upper and Additional Gypsum Units (UGU and AGU, respectively). Two cryptotephras were identified with glass analyses that correlate with tephra deposits from the Süphan and Nemrut volcanoes indicating that the AGU is ~1000 years younger than previously assumed, shifting it into the YD, and the underlying varved interval into the Bølling/Allerød, contradicting previous assumptions.
Using microfacies analyses, stable isotopes and temperature reconstructions, the second aim was achieved at Lake Gościąż. The YD lake system was dynamic, characterized by higher aquatic bioproductivity, more re-suspended material and less anoxia than during the Allerød and Early Holocene, mainly influenced by stronger water circulation and catchment erosion due to stronger westerly winds and less lake sheltering. Cooling at the YD onset was ~100 years longer than the final warming, while environmental proxies lagged the onset of cooling by ~90 years, but occurred contemporaneously during the termination of the YD. Chironomid-based temperature reconstructions support recent studies indicating mild YD summer temperatures. Such a comparison of annually-resolved proxy responses to both abrupt YD transitions is rare, because most European lake archives do not preserve varves during the YD.
To accomplish the second aim at the DS, microfacies analyses were performed between the UGU (~17 ka) and Holocene onset (~11 ka) in shallow- (Masada) and deep-water (ICDP core) environments. This time interval is marked by a huge but fluctuating lake level drop and therefore the complete transition into the Holocene is only recorded in the deep-basin ICDP core. In this thesis, this transition was investigated for the first time continuously and in detail. The final two pronounced lake level drops recorded by deposition of the UGU and AGU, were interrupted by one millennium of relative depositional stability and a positive water budget as recorded by aragonite varve deposition interrupted by only a few event layers. Further, intercalation of aragonite varves between the gypsum beds of the UGU and AGU shows that these generally dry intervals were also marked by decadal- to centennial-long rises in lake level. While continuous aragonite varves indicate decadal-long stable phases, the occurrence of thicker and more frequent event layers suggests general more instability during the gypsum units. These results suggest a pattern of complex and variable hydroclimate at different time scales during the Lateglacial at the DS.
The third aim was accomplished based on the individual studies above that jointly provide an integrated picture of different lake responses to different climate elements of hemispheric-scale abrupt climate changes during the Last Glacial-Interglacial transition. In general, climatically-driven facies changes are more dramatic in the DS than at Lake Gościąż. Further, Lake Gościąż is characterized by continuous varve formation nearly throughout the complete profile, whereas the DS record is widely characterized by extreme event layers, hampering the establishment of a continuous varve chronology. The lateglacial sedimentation in Lake Gościąż is mainly influenced by westerly winds and minor by changes in catchment vegetation, whereas the DS is primarily influenced by changes in winter precipitation, which are caused by temperature variations in the Mediterranean. Interestingly, sedimentation in both archives is more stable during the Bølling/Allerød and more dynamic during the YD, even when sedimentation processes are different.
In summary, this doctoral thesis presents seasonally-resolved records from two lake archives during the Lateglacial (ca 17-11 ka) to investigate the impact of abrupt climate changes in different lake systems. New age constrains from the identification of volcanic glass shards in the lateglacial sediments of the DS allowed the first lithology-based interpretation of the YD in the DS record and its comparison to Lake Gościąż. This highlights the importance of the construction of a robust chronology, and provides a first step for synchronization of the DS with other eastern Mediterranean archives. Further, climate reconstructions from the lake sediments showed variability on different time scales in the different archives, i.e. decadal- to millennial fluctuations in the lateglacial DS, and even annual variations and sub-decadal leads and lags in proxy responses during the rapid YD transitions in Lake Gościąż. This showed the importance of a comparison of different lake archives to better understand the regional and local impacts of hemispheric-scale climate variability. An unprecedented example is demonstrated here of how different lake systems show different lake responses and also react to different climate elements of abrupt climate changes. This further highlights the importance of the understanding of the respective lake system for climate reconstructions.
This study examines the access to healthcare for children and adolescents with three common chronic diseases (type-1 diabetes (T1D), obesity, or juvenile idiopathic arthritis (JIA)) within the 4th (Delta), 5th (Omicron), and beginning of the 6th (Omicron) wave (June 2021 until July 2022) of the COVID-19 pandemic in Germany in a cross-sectional study using three national patient registries. A paper-and-pencil questionnaire was given to parents of pediatric patients (<21 years) during the routine check-ups. The questionnaire contains self-constructed items assessing the frequency of healthcare appointments and cancellations, remote healthcare, and satisfaction with healthcare. In total, 905 parents participated in the T1D-sample, 175 in the obesity-sample, and 786 in the JIA-sample. In general, satisfaction with healthcare (scale: 0–10; 10 reflecting the highest satisfaction) was quite high (median values: T1D 10, JIA 10, obesity 8.5). The proportion of children and adolescents with canceled appointments was relatively small (T1D 14.1%, JIA 11.1%, obesity 20%), with a median of 1 missed appointment, respectively. Only a few parents (T1D 8.6%; obesity 13.1%; JIA 5%) reported obstacles regarding health services during the pandemic. To conclude, it seems that access to healthcare was largely preserved for children and adolescents with chronic health conditions during the COVID-19 pandemic in Germany.
Successful communication is often explored by people throughout their life courses. To effectively transfer one’s own information to others, people employ various linguistic tools, such as word order information, prosodic cues, and lexical choices. The exploration of these linguistic cues is known as the study of information structure (IS). Moreover, an important issue in the language acquisition of children is the investigation of how they acquire IS. This thesis seeks to improve our understanding of how children acquire different tools (i.e., prosodical cues, syntactical cues, and the focus particle only) of focus marking in a cross linguistic perspective.
In the first study, following Szendrői and her colleagues (2017)- the sentence-picture verification task- was performed to investigate whether three- to five-year-old Mandarin-speaking children as well as Mandarin-speaking adults could apply prosodic information to recognize focus in sentences. More, in the second study, not only Mandarin-speaking adults and Mandarin-speaking children but also German-speaking adults and German-speaking children were included to confirm the assumption that children could have adult-like performance in understanding sentence focus by identifying language specific cues in their mother tongue from early onwards. In this study, the same paradigm- the sentence-picture verification task- as in the first study was employed together with the eye-tracking method. Finally, in the last study, an issue of whether five-year-old Mandarin-speaking children could understand the pre-subject only sentence was carried out and again whether prosodic information would help them to better understand this kind of sentences.
The overall results seem to suggest that Mandarin-speaking children from early onwards could make use of the specific linguistic cues in their ambient language. That is, in Mandarin, a Topic-prominent and tone language, the word order information plays a more important rule than the prosodic information and even three-year-old Mandarin-speaking children could follow the word order information. More, although it seems that German-speaking children could follow the prosodic information, they did not have the adult-like performance in the object-accented condition. A feasible reason for this result is that there are more possibilities of marking focus in German, such as flexible word order, prosodic information, focus particles, and thus it would take longer time for German-speaking children to manage these linguistic tools. Another important empirical finding regarding the syntactically-marked focus in German is that it seems that the cleft construction is not a valid focus construction and this result corroborates with the previous observations (Dufter, 2009). Further, eye-tracking method did help to uncover how the parser direct their attention for recognizing focus. In the final study, it is showed that with explicit verbal context Mandarin-speaking children could understand the pre-subject only sentence and the study brought a better understanding of the acquisition of the focus particle- only with the Mandarin-speaking children.
Flood risk management in Germany follows an integrative approach in which both private households and businesses can make an important contribution to reducing flood damage by implementing property-level adaptation measures. While the flood adaptation behavior of private households has already been widely researched, comparatively less attention has been paid to the adaptation strategies of businesses. However, their ability to cope with flood risk plays an important role in the social and economic development of a flood-prone region. Therefore, using quantitative survey data, this study aims to identify different strategies and adaptation drivers of 557 businesses damaged by a riverine flood in 2013 and 104 businesses damaged by pluvial or flash floods between 2014 and 2017. Our results indicate that a low perceived self-efficacy may be an important factor that can reduce the motivation of businesses to adapt to flood risk. Furthermore, property-owners tended to act more proactively than tenants. In addition, high experience with previous flood events and low perceived response costs could strengthen proactive adaptation behavior. These findings should be considered in business-tailored risk communication.
Background
Wearables, as small portable computer systems worn on the body, can track user fitness and health data, which can be used to customize health insurance contributions individually. In particular, insured individuals with a healthy lifestyle can receive a reduction of their contributions to be paid. However, this potential is hardly used in practice.
Objective
This study aims to identify which barrier factors impede the usage of wearables for assessing individual risk scores for health insurances, despite its technological feasibility, and to rank these barriers according to their relevance.
Methods
To reach these goals, we conduct a ranking-type Delphi study with the following three stages. First, we collected possible barrier factors from a panel of 16 experts and consolidated them to a list of 11 barrier categories. Second, the panel was asked to rank them regarding their relevance. Third, to enhance the panel consensus, the ranking was revealed to the experts, who were then asked to re-rank the barriers.
Results
The results suggest that regulation is the most important barrier. Other relevant barriers are false or inaccurate measurements and application errors caused by the users. Additionally, insurers could lack the required technological competence to use the wearable data appropriately.
Conclusion
A wider use of wearables and health apps could be achieved through regulatory modifications, especially regarding privacy issues. Even after assuring stricter regulations, users’ privacy concerns could partly remain, if the data exchange between wearables manufacturers, health app providers, and health insurers does not become more transparent.
The present dissertation conducts empirical research on the relationship between urban life and its economic costs, especially for the environment. On the one hand, existing gaps in research on the influence of population density on air quality are closed and, on the other hand, innovative policy measures in the transport sector are examined that are intended to make metropolitan areas more sustainable. The focus is on air pollution, congestion and traffic accidents, which are important for general welfare issues and represent significant cost factors for urban life. They affect a significant proportion of the world's population. While 55% of the world's people already lived in cities in 2018, this share is expected to reach approximately 68% by 2050.
The four self-contained chapters of this thesis can be divided into two sections: Chapters 2 and 3 provide new causal insights into the complex interplay between urban structures and air pollution. Chapters 4 and 5 then examine policy measures to promote non-motorised transport and their influence on air quality as well as congestion and traffic accidents.
Wild bee species are important pollinators in agricultural landscapes. However, population decline was reported over the last decades and is still ongoing. While agricultural intensification is a major driver of the rapid loss of pollinating species, transition zones between arable fields and forest or grassland patches, i.e., agricultural buffer zones, are frequently mentioned as suitable mitigation measures to support wild bee populations and other pollinator species. Despite the reported general positive effect, it remains unclear which amount of buffer zones is needed to ensure a sustainable and permanent impact for enhancing bee diversity and abundance. To address this question at a pollinator community level, we implemented a process-based, spatially explicit simulation model of functional bee diversity dynamics in an agricultural landscape. More specifically, we introduced a variable amount of agricultural buffer zones (ABZs) at the transition of arable to grassland, or arable to forest patches to analyze the impact on bee functional diversity and functional richness. We focused our study on solitary bees in a typical agricultural area in the Northeast of Germany. Our results showed positive effects with at least 25% of virtually implemented agricultural buffer zones. However, higher amounts of ABZs of at least 75% should be considered to ensure a sufficient increase in Shannon diversity and decrease in quasi-extinction risks. These high amounts of ABZs represent effective conservation measures to safeguard the stability of pollination services provided by solitary bee species. As the model structure can be easily adapted to other mobile species in agricultural landscapes, our community approach offers the chance to compare the effectiveness of conservation measures also for other pollinator communities in future.
Algorithmic management
(2022)
Background
Eating in absence of hunger is quite common and often associated with an increased energy intake co-existent with a poorer food choice. Intuitive eating (IE), i.e., eating in accordance with internal hunger and satiety cues, may protect from overeating. IE, however, requires accurate perception and processing of one’s own bodily signals, also referred to as interoceptive sensitivity. Training interoceptive sensitivity might therefore be an effective method to promote IE and prevent overeating. As most studies on eating behavior are conducted in younger adults and close social relationships influence health-related behavior, this study focuses on middle-aged and older couples.
Methods
The present pilot randomized intervention study aims at investigating the feasibility and effectiveness of a 21-day mindfulness-based training program designed to increase interoceptive sensitivity. A total of N = 60 couples participating in the NutriAct Family Study, aged 50–80 years, will be recruited. This randomized-controlled intervention study comprises three measurement points (pre-intervention, post-intervention, 4-week follow-up) and a 21-day training that consists of daily mindfulness-based guided audio exercises (e.g., body scan). A three-arm intervention study design is applied to compare two intervention groups (training together as a couple vs. training alone) with a control group (no training). Each measurement point includes the assessment of self-reported and objective indicators of interoceptive sensitivity (primary outcome), self-reported indicators of intuitive and maladaptive eating (secondary outcomes), and additional variables. A training evaluation applying focus group discussions will be conducted to assess participants’ overall acceptance of the training and its feasibility.
Discussion
By investigating the feasibility and effectiveness of a mindfulness-based training program to increase interoceptive sensitivity, the present study will contribute to a deeper understanding of how to promote healthy eating in older age.
Physical activity and exercise are effective approaches in prevention and therapy of multiple diseases. Although the specific characteristics of lengthening contractions have the potential to be beneficial in many clinical conditions, eccentric training is not commonly used in clinical populations with metabolic, orthopaedic, or neurologic conditions. The purpose of this pilot study is to investigate the feasibility, functional benefits, and systemic responses of an eccentric exercise program focused on the trunk and lower extremities in people with low back pain (LBP) and multiple sclerosis (MS). A six-week eccentric training program with three weekly sessions is performed by people with LBP and MS. The program consists of ten exercises addressing strength of the trunk and lower extremities. The study follows a four-group design (N = 12 per group) in two study centers (Israel and Germany): three groups perform the eccentric training program: A) control group (healthy, asymptomatic); B) people with LBP; C) people with MS; group D (people with MS) receives standard care physiotherapy. Baseline measurements are conducted before first training, post-measurement takes place after the last session both comprise blood sampling, self-reported questionnaires, mobility, balance, and strength testing. The feasibility of the eccentric training program will be evaluated using quantitative and qualitative measures related to the study process, compliance and adherence, safety, and overall program assessment. For preliminary assessment of potential intervention effects, surrogate parameters related to mobility, postural control, muscle strength and systemic effects are assessed. The presented study will add knowledge regarding safety, feasibility, and initial effects of eccentric training in people with orthopaedic and neurological conditions. The simple exercises, that are easily modifiable in complexity and intensity, are likely beneficial to other populations. Thus, multiple applications and implementation pathways for the herein presented training program are conceivable.
Technological progress allows for producing ever more complex predictive models on the basis of increasingly big datasets. For risk management of natural hazards, a multitude of models is needed as basis for decision-making, e.g. in the evaluation of observational data, for the prediction of hazard scenarios, or for statistical estimates of expected damage. The question arises, how modern modelling approaches like machine learning or data-mining can be meaningfully deployed in this thematic field. In addition, with respect to data availability and accessibility, the trend is towards open data. Topic of this thesis is therefore to investigate the possibilities and limitations of machine learning and open geospatial data in the field of flood risk modelling in the broad sense. As this overarching topic is broad in scope, individual relevant aspects are identified and inspected in detail.
A prominent data source in the flood context is satellite-based mapping of inundated areas, for example made openly available by the Copernicus service of the European Union. Great expectations are directed towards these products in scientific literature, both for acute support of relief forces during emergency response action, and for modelling via hydrodynamic models or for damage estimation. Therefore, a focus of this work was set on evaluating these flood masks. From the observation that the quality of these products is insufficient in forested and built-up areas, a procedure for subsequent improvement via machine learning was developed. This procedure is based on a classification algorithm that only requires training data from a particular class to be predicted, in this specific case data of flooded areas, but not of the negative class (dry areas). The application for hurricane Harvey in Houston shows the high potential of this method, which depends on the quality of the initial flood mask.
Next, it is investigated how much the predicted statistical risk from a process-based model chain is dependent on implemented physical process details. Thereby it is demonstrated what a risk study based on established models can deliver. Even for fluvial flooding, such model chains are already quite complex, though, and are hardly available for compound or cascading events comprising torrential rainfall, flash floods, and other processes. In the fourth chapter of this thesis it is therefore tested whether machine learning based on comprehensive damage data can offer a more direct path towards damage modelling, that avoids explicit conception of such a model chain. For that purpose, a state-collected dataset of damaged buildings from the severe El Niño event 2017 in Peru is used. In this context, the possibilities of data-mining for extracting process knowledge are explored as well. It can be shown that various openly available geodata sources contain useful information for flood hazard and damage modelling for complex events, e.g. satellite-based rainfall measurements, topographic and hydrographic information, mapped settlement areas, as well as indicators from spectral data. Further, insights on damaging processes are discovered, which mainly are in line with prior expectations. The maximum intensity of rainfall, for example, acts stronger in cities and steep canyons, while the sum of rain was found more informative in low-lying river catchments and forested areas. Rural areas of Peru exhibited higher vulnerability in the presented study compared to urban areas. However, the general limitations of the methods and the dependence on specific datasets and algorithms also become obvious.
In the overarching discussion, the different methods – process-based modelling, predictive machine learning, and data-mining – are evaluated with respect to the overall research questions. In the case of hazard observation it seems that a focus on novel algorithms makes sense for future research. In the subtopic of hazard modelling, especially for river floods, the improvement of physical models and the integration of process-based and statistical procedures is suggested. For damage modelling the large and representative datasets necessary for the broad application of machine learning are still lacking. Therefore, the improvement of the data basis in the field of damage is currently regarded as more important than the selection of algorithms.
The development of speaking competence is widely regarded as a central aspect of second language (L2) learning. It may be questioned, however, if the currently predominant ways of conceptualising the term fully satisfy the complexity of the construct: Although there is growing recognition that language primarily constitutes a tool for communication and participation in social life, as yet it is rare for conceptualisations of speaking competence to incorporate the ability to inter-act and co-construct meaning with co-participants. Accordingly, skills allowing for the successful accomplishment of interactional tasks (such as orderly speaker change, and resolving hearing and understanding trouble) also remain largely unrepresented in language teaching and assessment. As fostering the ability to successfully use the L2 within social interaction should arguably be a main objective of language teaching, it appears pertinent to broaden the construct of speaking competence by incorporating interactional competence (IC). Despite there being a growing research interest in the conceptualisation and development of (L2) IC, much of the materials and instruments required for its teaching and assessment, and thus for fostering a broader understanding of speaking competence in the L2 classroom, still await development. This book introduces an approach to the identification of candidate criterial features for the assessment of EFL learners’ L2 repair skills. Based on a corpus of video-recorded interaction between EFL learners, and following conversation-analytic and interactional-linguistic methodology as well as drawing on basic premises of research in the framework of Conversation Analysis for Second Language Acquisition, differences between (groups of) learners in terms of their L2 repair conduct are investigated through qualitative and inductive analyses. Candidate criterial features are derived from the analysis results. This book does not only contribute to the operationalisation of L2 IC (and of L2 repair skills in particular), but also lays groundwork for the construction of assessment scales and rubrics geared towards the evaluation of EFL learners’ L2 interactional skills.
Despite the positive effects of including patients’ preferences into therapy on psychotherapy outcomes, there are still few thoroughly validated assessment tools at hand. We translated the 18-item Cooper-Norcross Inventory of Preferences (C-NIP) into German and aimed at replicating its factor structure. Further, we investigated the reliability of the questionnaire and its convergence with trait measures. A heterogeneous sample of N = 969 participants took part in our online survey. Performing ESEM models, we found acceptable model fit for a four-factor structure similar to the original factor structure. Furthermore, we propose an alternative model following the adjustment of single items. The German C-NIP showed acceptable to good reliability, as well as small correlations with Big-Five personality traits, trait and attachment anxiety, locus of control, and temporal focus. However, we recommend further replication of the factor structure and further validation of the C-NIP.
Cosmic-ray neutron sensing (CRNS) is a non-invasive tool for measuring hydrogen pools such as soil moisture, snow or vegetation. The intrinsic integration over a radial hectare-scale footprint is a clear advantage for averaging out small-scale heterogeneity, but on the other hand the data may become hard to interpret in complex terrain with patchy land use.
This study presents a directional shielding approach to prevent neutrons from certain angles from being counted while counting neutrons entering the detector from other angles and explores its potential to gain a sharper horizontal view on the surrounding soil moisture distribution.
Using the Monte Carlo code URANOS (Ultra Rapid Neutron-Only Simulation), we modelled the effect of additional polyethylene shields on the horizontal field of view and assessed its impact on the epithermal count rate, propagated uncertainties and aggregation time.
The results demonstrate that directional CRNS measurements are strongly dominated by isotropic neutron transport, which dilutes the signal of the targeted direction especially from the far field. For typical count rates of customary CRNS stations, directional shielding of half-spaces could not lead to acceptable precision at a daily time resolution. However, the mere statistical distinction of two rates should be feasible.
Environmental pollution by microplastics has become a severe problem in terrestrial and aquatic ecosystems and, according to actual prognoses, problems will further increase in the future. Therefore, assessing and quantifying the risk for the biota is crucial. Standardized short-term toxicological procedures as well as methods quantifying potential toxic effects over the whole life span of an animal are required. We studied the effect of the microplastic polystyrene on the survival and reproduction of a common freshwater invertebrate, the rotifer Brachionus calyciflorus, at different timescales. We used pristine polystyrene spheres of 1, 3, and 6 µm diameter and fed them to the animals together with food algae in different ratios ranging from 0 to 50% nonfood particles. As a particle control, we used silica to distinguish between a pure particle effect and a plastic effect. After 24 h, no toxic effect was found, neither with polystyrene nor with silica. After 96 h, a toxic effect was detectable for both particle types. The size of the particles played a negligible role. Studying the long-term effect by using life table experiments, we found a reduced reproduction when the animals were fed with 3 µm spheres together with similar-sized food algae. We conclude that the fitness reduction is mainly driven by the dilution of food by the nonfood particles rather than by a direct toxic effect.
Background and Aims Wearable inertial sensors may offer additional kinematic parameters of the shoulder compared to traditional instruments such as goniometers when elaborate and time-consuming data processing procedures are undertaken. However, in clinical practice simple-real time motion analysis is required to improve clinical reasoning. Therefore, the aim was to assess the criterion validity between a portable "off-the-shelf" sensor-software system (IMU) and optical motion (Mocap) for measuring kinematic parameters during active shoulder movements. Methods 24 healthy participants (9 female, 15 male, age 29 +/- 4 years, height 177 +/- 11 cm, weight 73 +/- 14 kg) were included. Range of motion (ROM), total range of motion (TROM), peak and mean angular velocity of both systems were assessed during simple (abduction/adduction, horizontal flexion/horizontal extension, vertical flexion/extension, and external/internal rotation) and complex shoulder movements. Criterion validity was determined using intraclass-correlation coefficients (ICC), root mean square error (RMSE) and Bland and Altmann analysis (bias; upper and lower limits of agreement). Results ROM and TROM analysis revealed inconsistent validity during simple (ICC: 0.040-0.733, RMSE: 9.7 degrees-20.3 degrees, bias: 1.2 degrees-50.7 degrees) and insufficient agreement during complex shoulder movements (ICC: 0.104-0.453, RMSE: 10.1 degrees-23.3 degrees, bias: 1.0 degrees-55.9 degrees). Peak angular velocity (ICC: 0.202-0.865, RMSE: 14.6 degrees/s-26.7 degrees/s, bias: 10.2 degrees/s-29.9 degrees/s) and mean angular velocity (ICC: 0.019-0.786, RMSE:6.1 degrees/s-34.2 degrees/s, bias: 1.6 degrees/s-27.8 degrees/s) were inconsistent. Conclusions The "off-the-shelf" sensor-software system showed overall insufficient agreement with the gold standard. Further development of commercial IMU-software-solutions may increase measurement accuracy and permit their integration into everyday clinical practice.
Association of primary allostatic load mediators and metabolic syndrome (MetS): A systematic review
(2022)
Allostatic load (AL) exposure may cause detrimental effects on the neuroendocrine system, leading to metabolic syndrome (MetS). The primary mediators of AL involve serum dehydroepiandrosterone sulfate (DHEAS; a functional HPA axis antagonist); further, cortisol, urinary norepinephrine (NE), and epinephrine (EPI) excretion levels (assessed within 12-h urine as a golden standard for the evaluation of the HPA axis activity and sympathetic nervous system activity). However, the evidence of an association between the primary mediators of AL and MetS is limited. This systematic review aimed to critically examine the association between the primary mediators of AL and MetS. PubMed and Web of Science were searched for articles from January 2010 to December 2021, published in English. The search strategy focused on cross-sectional and case–control studies comprising adult participants with MetS, obesity, overweight, and without chronic diseases. The STROBE checklist was used to assess study quality control. Of 770 studies, twenty-one studies with a total sample size (n = 10,666) met the eligibility criteria. Eighteen studies were cross-sectional, and three were case–control studies. The included studies had a completeness of reporting score of COR % = 87.0 ± 6.4%. It is to be noted, that cortisol as a primary mediator of AL showed an association with MetS in 50% (urinary cortisol), 40% (serum cortisol), 60% (salivary cortisol), and 100% (hair cortisol) of the studies. For DHEAS, it is to conclude that 60% of the studies showed an association with MetS. In contrast, urinary EPI and urinary NE had 100% no association with MetS. In summary, there is a tendency for the association between higher serum cortisol, salivary cortisol, urinary cortisol, hair cortisol, and lower levels of DHEAS with MetS. Future studies focusing on longitudinal data are warranted for clarification and understanding of the association between the primary mediators of AL and MetS.
Cyberhate represents a risk to adolescents’ development and peaceful coexistence in democratic societies. Yet, not much is known about the relationship between adolescents’ ability to cope with cyberhate and their cyberhate involvement. To fill current gaps in the literature and inform the development of media education programs, the present study investigated various coping strategies in a hypothetical cyberhate scenario as correlates for being cyberhate victims, perpetrators, and both victim–perpetrators. The sample consisted of 6829 adolescents aged 12–18 years old (Mage = 14.93, SD = 1.64; girls: 50.4%, boys: 48.9%, and 0.7% did not indicate their gender) from Asia, Europe, and North America. Results showed that adolescents who endorsed distal advice or endorsed technical coping showed a lower likelihood to be victims, perpetrators, or victim–perpetrators. In contrast, if adolescents felt helpless or endorsed retaliation to cope with cyberhate, they showed higher odds of being involved in cyberhate as victims, perpetrators, or victim–perpetrators. Finally, adolescents who endorsed close support as a coping strategy showed a lower likelihood to be victim–perpetrators, and adolescents who endorsed assertive coping showed higher odds of being victims. In conclusion, the results confirm the importance of addressing adolescents’ ability to deal with cyberhate to develop more tailored prevention approaches. More specifically, such initiatives should focus on adolescents who feel helpless or feel inclined to retaliate. In addition, adolescents should be educated to practice distal advice and technical coping when experiencing cyberhate. Implications for the design and instruction of evidence-based cyberhate prevention (e.g., online educational games, virtual learning environments) will be discussed.
A decade ago, it became feasible to store multi-terabyte databases in main memory. These in-memory databases (IMDBs) profit from DRAM's low latency and high throughput as well as from the removal of costly abstractions used in disk-based systems, such as the buffer cache. However, as the DRAM technology approaches physical limits, scaling these databases becomes difficult. Non-volatile memory (NVM) addresses this challenge. This new type of memory is persistent, has more capacity than DRAM (4x), and does not suffer from its density-inhibiting limitations. Yet, as NVM has a higher latency (5-15x) and a lower throughput (0.35x), it cannot fully replace DRAM.
IMDBs thus need to navigate the trade-off between the two memory tiers. We present a solution to this optimization problem. Leveraging information about access frequencies and patterns, our solution utilizes NVM's additional capacity while minimizing the associated access costs. Unlike buffer cache-based implementations, our tiering abstraction does not add any costs when reading data from DRAM. As such, it can act as a drop-in replacement for existing IMDBs. Our contributions are as follows:
(1) As the foundation for our research, we present Hyrise, an open-source, columnar IMDB that we re-engineered and re-wrote from scratch. Hyrise enables realistic end-to-end benchmarks of SQL workloads and offers query performance which is competitive with other research and commercial systems. At the same time, Hyrise is easy to understand and modify as repeatedly demonstrated by its uses in research and teaching.
(2) We present a novel memory management framework for different memory and storage tiers. By encapsulating the allocation and access methods of these tiers, we enable existing data structures to be stored on different tiers with no modifications to their implementation. Besides DRAM and NVM, we also support and evaluate SSDs and have made provisions for upcoming technologies such as disaggregated memory.
(3) To identify the parts of the data that can be moved to (s)lower tiers with little performance impact, we present a tracking method that identifies access skew both in the row and column dimensions and that detects patterns within consecutive accesses. Unlike existing methods that have substantial associated costs, our access counters exhibit no identifiable overhead in standard benchmarks despite their increased accuracy.
(4) Finally, we introduce a tiering algorithm that optimizes the data placement for a given memory budget. In the TPC-H benchmark, this allows us to move 90% of the data to NVM while the throughput is reduced by only 10.8% and the query latency is increased by 11.6%. With this, we outperform approaches that ignore the workload's access skew and access patterns and increase the query latency by 20% or more.
Individually, our contributions provide novel approaches to current challenges in systems engineering and database research. Combining them allows IMDBs to scale past the limits of DRAM while continuing to profit from the benefits of in-memory computing.
Basic psychological needs theory postulates that a social environment that satisfies individuals’ three basic psychological needs of autonomy, competence, and relatedness leads to optimal growth and well-being. On the other hand, the frustration of these needs is associated with ill-being and depressive symptoms foremost investigated in non-clinical samples; yet, there is a paucity of research on need frustration in clinical samples. Survey data were compared between adult individuals with major depressive disorder (MDD; n = 115; 48.69% female; 38.46 years, SD = 10.46) with those of a non-depressed comparison sample (n = 201; 53.23% female; 30.16 years, SD = 12.81). Need profiles were examined with a linear mixed model (LMM). Individuals with depression reported higher levels of frustration and lower levels of satisfaction in relation to the three basic psychological needs when compared to non-depressed adults. The difference between depressed and non-depressed groups was significantly larger for frustration than satisfaction regarding the needs for relatedness and competence. LMM correlation parameters confirmed the expected positive correlation between the three needs. This is the first study showing substantial differences in need-based experiences between depressed and non-depressed adults. The results confirm basic assumptions of the self-determination theory and have preliminary implications in tailoring therapy for depression.
Sprache
Englisch
Modern single-particle-tracking techniques produce extensive time-series of diffusive motion in a wide variety of systems, from single-molecule motion in living-cells to movement ecology. The quest is to decipher the physical mechanisms encoded in the data and thus to better understand the probed systems. We here augment recently proposed machine-learning techniques for decoding anomalous-diffusion data to include an uncertainty estimate in addition to the predicted output. To avoid the Black-Box-Problem a Bayesian-Deep-Learning technique named Stochastic-Weight-Averaging-Gaussian is used to train models for both the classification of the diffusionmodel and the regression of the anomalous diffusion exponent of single-particle-trajectories. Evaluating their performance, we find that these models can achieve a wellcalibrated error estimate while maintaining high prediction accuracies. In the analysis of the output uncertainty predictions we relate these to properties of the underlying diffusion models, thus providing insights into the learning process of the machine and the relevance of the output.
Predation is a strong species interaction causing severe harm or death to prey. Thus, prey species have evolved various defence strategies to minimize predation risk, which may be immediate (e.g., a change in behaviour) or transgenerational (morphological defence structures). We studied the behaviour of two strains of a rotiferan prey (Brachionus calyciflorus) that differ in their ability to develop morphological defences in response to their predator Asplanchna brightwellii. Using video analysis, we tested: (a) if two strains differ in their response to predator presence and predator cues when both are undefended; (b) whether defended individuals respond to live predators or their cues; and (c) if the morphological defence (large spines) per se has an effect on the swimming behaviour. We found a clear increase in swimming speed for both undefended strains in predator presence. However, the defended specimens responded neither to the predator presence nor to their cues, showing that they behave indifferently to their predator when they are defended. We did not detect an effect of the spines on the swimming behaviour. Our study demonstrates a complex plastic behaviour of the prey, not only in the presence of their predator, but also with respect to their defence status.
There is broad agreement among researchers to view mind wandering as an obstacle to learning because it draws attention away from learning tasks. Accordingly, empirical findings revealed negative correlations between the frequency of mind wandering during learning and various kinds of learning outcomes (e.g., text retention). However, a few studies have indicated positive effects of mind wandering on creativity in real-world learning environments. The present article reviews these studies and highlights potential benefits of mind wandering for learning mediated through creative processes. Furthermore, we propose various ways to promote useful mind wandering and, at the same time, minimize its negative impact on learning.
This study examined the spoken narrative skills of a group of bilingual Mandarin–English speaking 3–6-year-olds (N = 25) in Australia, using a remote online story-retell task. Bilingual preschoolers are an understudied population, especially those who are speaking typologically distinct languages such as Mandarin and English which have fewer structural overlaps compared to language pairs that are typologically closer, reducing cross-linguistic positive transfer. We examined these preschoolers’ spoken narrative skills as measured by macrostructures (the global organization of a story) and microstructures (linguistic structures, e.g., total number of utterances, nouns, verbs, phrases, and modifiers) across and within each language, and how various factors such as age and language experiences contribute to individual variability. The results indicate that our bilingual preschoolers acquired spoken narrative skills similarly across their two languages, i.e., showing similar patterns of productivity for macrostructure and microstructure elements in both of their two languages. While chronological age was positively correlated with macrostructures in both languages (showing developmental effects), there were no significant correlations between measures of language experiences and the measures of spoken narrative skills (no effects for language input/output). The findings suggest that although these preschoolers acquire two typologically diverse languages in different learning environments, Mandarin at home with highly educated parents, and English at preschool, they displayed similar levels of oral narrative skills as far as these macro−/micro-structure measures are concerned. This study provides further evidence for the feasibility of remote online assessment of preschoolers’ narrative skills.
Bio-sourced adsorbing poly(2-oxazoline)s mimicking mussel glue proteins for antifouling applications
(2022)
Nature developed countless systems for many applications. In maritime environments, several organisms established extra-ordinary mechanisms to attach to surfaces. Over the past years, the scientific interest to employ those mechanisms for coatings and long-lasting adhering materials gained significant attention.
This work describes the synthesis of bio-inspired adsorbing copoly(2-oxazoline)s for surface coatings with protein repelling effects, mimicking mussel glue proteins. From a set of methoxy substituted phenyl, benzyl, and cinnamyl acids, 2-oxazoline monomers were synthesized. All synthesized 2-oxazolines were analyzed by FT-IR spectroscopy, NMR spectroscopy, and EI mass spectrometry. With those newly synthesized 2-oxazoline monomers and 2-ethyl-2-oxazoline, kinetic studies concerning homo- and copolymerization in a microwave reactor were conducted. The success of the polymerization reactions was demonstrated by FT-IR spectroscopy, NMR spectroscopy, MALDI-TOF mass spectrometry, and size exclusion chromatography (SEC). The copolymerization of 2-ethyl-2-oxazoline with a selection of methoxy-substituted 2-oxazolines resulted in water-soluble copolymers. To release the adsorbing catechol and cationic units, the copoly(2-oxazoline)s were modified. The catechol units were (partially) released by a methyl aryl ether cleavage reaction. A subsequent partial acidic hydrolysis of the ethyl unit resulted in mussel glue protein-inspired catechol and cation-containing copolymers. The modified copolymers were analyzed by NMR spectroscopy, UV-VIS spectroscopy, and SEC. The catechol- and cation-containing copolymers and their precursors were examined by a Quartz Crystal Microbalance with Dissipation (QCM-D), so study the adsorption performance on gold, borosilicate, iron, and polystyrene surfaces. An exemplary study revealed that a catechol and cation-containing copoly(2-oxazoline)-coated gold surface exhibits strong protein repelling properties.
Carbohydrates are found in every living organism, where they are responsible for numerous, essential biological functions and processes. Synthetic polymers with pendant saccharides, called glycopolymers, mimic natural glycoconjugates in their special properties and functions. Employing such biomimetics furthers the understanding and controlling of biological processes. Hence, glycopolymers are valuable and interesting for applications in the medical and biological field. However, the synthesis of carbohydrate-based materials can be very challenging. In this thesis, the synthesis of biofunctional glycopolymers is presented, with the focus on aqueous-based, protecting group free and short synthesis routes to further advance in the field of glycopolymer synthesis.
A practical and versatile precursor for glycopolymers are glycosylamines. To maintain biofunctionality of the saccharides after their amination, regioselective functionalization was performed. This frequently performed synthesis was optimized for different sugars. The optimization was facilitated using a design of experiment (DoE) approach to enable a reduced number of necessary experiments and efficient procedure. Here, the utility of using DoE for optimizing the synthesis of glycosylamines is discussed.
The glycosylamines were converted to glycomonomers which were then polymerized to yield biofunctional glycopolymers. Here, the glycopolymers were aimed to be applicable as layer-by-layer (LbL) thin film coatings for drug delivery systems. To enable the LbL technique, complimentary glycopolymer electrolytes were synthesized by polymerization of the glycomonomers and subsequent modification or by post-polymerization modification. For drug delivery, liposomes were embedded into the glycopolymer coating as potential cargo carriers. The stability as well as the integrity of the glycopolymer layers and liposomes were investigated at physiological pH range.
Different glycopolymers were also synthesized to be applicable as anti-adhesion therapeutics by providing advanced architectures with multivalent presentations of saccharides, which can inhibit the binding of pathogene lectins. Here, the synthesis of glycopolymer hydrogel particles based on biocompatible poly(N-isopropylacrylamide) (NiPAm) was established using the free-radical precipitation polymerization technique. The influence of synthesis parameters on the sugar content in the gels and on the hydrogel morphology is discussed. The accessibility of the saccharides to model lectins and their enhanced, multivalent interaction were investigated.
At the end of this work, the synthesis strategies for the glycopolymers are generally discussed as well as their potential application in medicine.
The index theorem for elliptic operators on a closed Riemannian manifold by Atiyah and Singer has many applications in analysis, geometry and topology, but it is not suitable for a generalization to a Lorentzian setting.
In the case where a boundary is present Atiyah, Patodi and Singer provide an index theorem for compact Riemannian manifolds by introducing non-local boundary conditions obtained via the spectral decomposition of an induced boundary operator, so called APS boundary conditions. Bär and Strohmaier prove a Lorentzian version of this index theorem for the Dirac operator on a manifold with boundary by utilizing results from APS and the characterization of the spectral flow by Phillips. In their case the Lorentzian manifold is assumed to be globally hyperbolic and spatially compact, and the induced boundary operator is given by the Riemannian Dirac operator on a spacelike Cauchy hypersurface. Their results show that imposing APS boundary conditions for these boundary operator will yield a Fredholm operator with a smooth kernel and its index can be calculated by a formula similar to the Riemannian case.
Back in the Riemannian setting, Bär and Ballmann provide an analysis of the most general kind of boundary conditions that can be imposed on a first order elliptic differential operator that will still yield regularity for solutions as well as Fredholm property for the resulting operator. These boundary conditions can be thought of as deformations to the graph of a suitable operator mapping APS boundary conditions to their orthogonal complement.
This thesis aims at applying the boundary conditions found by Bär and Ballmann to a Lorentzian setting to understand more general types of boundary conditions for the Dirac operator, conserving Fredholm property as well as providing regularity results and relative index formulas for the resulting operators. As it turns out, there are some differences in applying these graph-type boundary conditions to the Lorentzian Dirac operator when compared to the Riemannian setting. It will be shown that in contrast to the Riemannian case, going from a Fredholm boundary condition to its orthogonal complement works out fine in the Lorentzian setting. On the other hand, in order to deduce Fredholm property and regularity of solutions for graph-type boundary conditions, additional assumptions for the deformation maps need to be made.
The thesis is organized as follows. In chapter 1 basic facts about Lorentzian and Riemannian spin manifolds, their spinor bundles and the Dirac operator are listed. These will serve as a foundation to define the setting and prove the results of later chapters.
Chapter 2 defines the general notion of boundary conditions for the Dirac operator used in this thesis and introduces the APS boundary conditions as well as their graph type deformations. Also the role of the wave evolution operator in finding Fredholm boundary conditions is analyzed and these boundary conditions are connected to notion of Fredholm pairs in a given Hilbert space.
Chapter 3 focuses on the principal symbol calculation of the wave evolution operator and the results are used to proof Fredholm property as well as regularity of solutions for suitable graph-type boundary conditions. Also sufficient conditions are derived for (pseudo-)local boundary conditions imposed on the Dirac operator to yield a Fredholm operator with a smooth solution space.
In the last chapter 4, a few examples of boundary conditions are calculated applying the results of previous chapters. Restricting to special geometries and/or boundary conditions, results can be obtained that are not covered by the more general statements, and it is shown that so-called transmission conditions behave very differently than in the Riemannian setting.
Plant metabolism is the main process of converting assimilated carbon to different crucial compounds for plant growth and therefore crop yield, which makes it an important research topic. Although major advances in understanding genetic principles contributing to metabolism and yield have been made, little is known about the genetics responsible for trait variation or canalization although the concepts have been known for a long time. In light of a growing global population and progressing climate change, understanding canalization of metabolism and yield seems ever-more important to ensure food security. Our group has recently found canalization metabolite quantitative trait loci (cmQTL) for tomato fruit metabolism, showing that the concept of canalization applies on metabolism. In this work two approaches to investigate plant metabolic canalization and one approach to investigate yield canalization are presented.
In the first project, primary and secondary metabolic data from Arabidopsis thaliana and Phaseolus vulgaris leaf material, obtained from plants grown under different conditions was used to calculate cross-environment coefficient of variations or fold-changes of metabolite levels per genotype and used as input for genome wide association studies. While primary metabolites have lower CV across conditions and show few and mostly weak associations to genomic regions, secondary metabolites have higher CV and show more, strong metabolite to genome associations. As candidate genes, both potential regulatory genes as well as metabolic genes, can be found, albeit most metabolic genes are rarely directly related to the target metabolites, suggesting a role for both potential regulatory mechanisms as well as metabolic network structure for canalization of metabolism.
In the second project, candidate genes of the Solanum lycopersicum cmQTL mapping are selected and CRISPR/Cas9-mediated gene-edited tomato lines are created, to validate the genes role in canalization of metabolism. Obtained mutants appeared to either have strong aberrant developmental phenotypes or appear wild type-like. One phenotypically inconspicuous mutant of a pantothenate kinase, selected as candidate for malic acid canalization shows a significant increase of CV across different watering conditions. Another such mutant of a protein putatively involved in amino acid transport, selected as candidate for phenylalanine canalization shows a similar tendency to increased CV without statistical significance. This potential role of two genes involved in metabolism supports the hypothesis of structural relevance of metabolism for its own stability.
In the third project, a mutant for a putative disulfide isomerase, important for thylakoid biogenesis, is characterized by a multi-omics approach. The mutant was characterized previously in a yield stability screening and showed a variegated leaf phenotype, ranging from green leaves with wild type levels of chlorophyll over differently patterned variegated to completely white leaves almost completely devoid of photosynthetic pigments. White mutant leaves show wild type transcript levels of photosystem assembly factors, with the exception of ELIP and DEG orthologs indicating a stagnation at an etioplast to chloroplast transition state. Green mutant leaves show an upregulation of these assembly factors, possibly acting as overcompensation for partially defective disulfide isomerase, which seems sufficient for proper chloroplast development as confirmed by a wild type-like proteome. Likely as a result of this phenotype, a general stress response, a shift to a sink-like tissue and abnormal thylakoid membranes, strongly alter the metabolic profile of white mutant leaves. As the severity and pattern of variegation varies from plant to plant and may be effected by external factors, the effect on yield instability, may be a cause of a decanalized ability to fully exploit the whole leaf surface area for photosynthetic activity.
Digital transformation (DT) has not only been a major challenge in recent years, it is also supposed to continue to enormously impact our society and economy in the forthcoming decade. On the one hand, digital technologies have emerged, diffusing and determining our private and professional lives. On the other hand, digital platforms have leveraged the potentials of digital technologies to provide new business models. These dynamics have a massive effect on individuals, companies, and entire ecosystems. Digital technologies and platforms have changed the way persons consume or interact with each other. Moreover, they offer companies new opportunities to conduct their business in terms of value creation (e.g., business processes), value proposition (e.g., business models), or customer interaction (e.g., communication channels), i.e., the three dimensions of DT. However, they also can become a threat for a company's competitiveness or even survival. Eventually, the emergence, diffusion, and employment of digital technologies and platforms bear the potential to transform entire markets and ecosystems.
Against this background, IS research has explored and theorized the phenomena in the context of DT in the past decade, but not to its full extent. This is not surprising, given the complexity and pervasiveness of DT, which still requires far more research to further understand DT with its interdependencies in its entirety and in greater detail, particularly through the IS perspective at the confluence of technology, economy, and society. Consequently, the IS research discipline has determined and emphasized several relevant research gaps for exploring and understanding DT, including empirical data, theories as well as knowledge of the dynamic and transformative capabilities of digital technologies and platforms for both organizations and entire industries.
Hence, this thesis aims to address these research gaps on the IS research agenda and consists of two streams. The first stream of this thesis includes four papers that investigate the impact of digital technologies on organizations. In particular, these papers study the effects of new technologies on firms (paper II.1) and their innovative capabilities (II.2), the nature and characteristics of data-driven business models (II.3), and current developments in research and practice regarding on-demand healthcare (II.4). Consequently, the papers provide novel insights on the dynamic capabilities of digital technologies along the three dimensions of DT. Furthermore, they offer companies some opportunities to systematically explore, employ, and evaluate digital technologies to modify or redesign their organizations or business models.
The second stream comprises three papers that explore and theorize the impact of digital platforms on traditional companies, markets, and the economy and society at large. At this, paper III.1 examines the implications for the business of traditional insurance companies through the emergence and diffusion of multi-sided platforms, particularly in terms of value creation, value proposition, and customer interaction. Paper III.2 approaches the platform impact more holistically and investigates how the ongoing digital transformation and "platformization" in healthcare lastingly transform value creation in the healthcare market. Paper III.3 moves on from the level of single businesses or markets to the regulatory problems that result from the platform economy for economy and society, and proposes appropriate regulatory approaches for addressing these problems. Hence, these papers bring new insights on the table about the transformative capabilities of digital platforms for incumbent companies in particular and entire ecosystems in general.
Altogether, this thesis contributes to the understanding of the impact of DT on organizations and markets through the conduction of multiple-case study analyses that are systematically reflected with the current state of the art in research. On this empirical basis, the thesis also provides conceptual models, taxonomies, and frameworks that help describing, explaining, or predicting the impact of digital technologies and digital platforms on companies, markets and the economy or society at large from an interdisciplinary viewpoint.
Inter-brain synchronization is primarily investigated during social interactions but had not been examined during coupled muscle action between two persons until now. It was previously shown that mechanical muscle oscillations can develop coherent behavior between two isometrically interacting persons. This case study investigated if inter-brain synchronization appears thereby, and if differences of inter- and intrapersonal muscle and brain coherence exist regarding two different types of isometric muscle action. Electroencephalography (EEG) and mechanomyography/mechanotendography (MMG/MTG) of right elbow extensors were recorded during six fatiguing trials of two coupled isometrically interacting participants (70% MVIC). One partner performed holding and one pushing isometric muscle action (HIMA/PIMA; tasks changed). The wavelet coherence of all signals (EEG, MMG/MTG, force, ACC) were analyzed intra- and interpersonally. The five longest coherence patches in 8–15 Hz and their weighted frequency were compared between real vs. random pairs and between HIMA vs. PIMA. Real vs. random pairs showed significantly higher coherence for intra-muscle, intra-brain, and inter-muscle-brain activity (p < 0.001 to 0.019). Inter-brain coherence was significantly higher for real vs. random pairs for EEG of right and central areas and for sub-regions of EEG left (p = 0.002 to 0.025). Interpersonal muscle-brain synchronization was significantly higher than intrapersonal one, whereby it was significantly higher for HIMA vs. PIMA. These preliminary findings indicate that inter-brain synchronization can arise during muscular interaction. It is hypothesized both partners merge into one oscillating neuromuscular system. The results reinforce the hypothesis that HIMA is characterized by more complex control strategies than PIMA. The pilot study suggests investigating the topic further to verify these results on a larger sample size. Findings could contribute to the basic understanding of motor control and is relevant for functional diagnostics such as the manual muscle test which is applied in several disciplines, e.g., neurology, physiotherapy.
Previous research demonstrated a close bidirectional relationship between spatial attention and the manual motor system. However, it is unclear whether an explicit hand movement is necessary for this relationship to appear. A novel method with high temporal resolution–bimanual grip force registration–sheds light on this issue. Participants held two grip force sensors while being presented with lateralized stimuli (exogenous attentional shifts, Experiment 1), left- or right-pointing central arrows (endogenous attentional shifts, Experiment 2), or the words "left" or "right" (endogenous attentional shifts, Experiment 3). There was an early interaction between the presentation side or arrow direction and grip force: lateralized objects and central arrows led to a larger increase of the ipsilateral force and a smaller increase of the contralateral force. Surprisingly, words led to the opposite pattern: larger force increase in the contralateral hand and smaller force increase in the ipsilateral hand. The effect was stronger and appeared earlier for lateralized objects (60 ms after stimulus presentation) than for arrows (100 ms) or words (250 ms). Thus, processing visuospatial information automatically activates the manual motor system, but the timing and direction of this effect vary depending on the type of stimulus.
Over the past decades, there has been a growing interest in ‘extreme events’ owing to the increasing threats that climate-related extremes such as floods, heatwaves, droughts, etc., pose to society. While extreme events have diverse definitions across various disciplines, ranging from earth science to neuroscience, they are characterized mainly as dynamic occurrences within a limited time frame that impedes the normal functioning of a system. Although extreme events are rare in occurrence, it has been found in various hydro-meteorological and physiological time series (e.g., river flows, temperatures, heartbeat intervals) that they may exhibit recurrent behavior, i.e., do not end the lifetime of the system. The aim of this thesis to develop some
sophisticated methods to study various properties of extreme events.
One of the main challenges in analyzing such extreme event-like time series is that they have large temporal gaps due to the paucity of the number of observations of extreme events. As a result, existing time series analysis tools are usually not helpful to decode the underlying
information. I use the edit distance (ED) method to analyze extreme event-like time series in their unaltered form. ED is a specific distance metric, mainly designed to measure the similarity/dissimilarity between point process-like data. I combine ED with recurrence plot techniques to identify the recurrence property of flood events in the Mississippi River in the United States. I also use recurrence quantification analysis to show the deterministic properties
and serial dependency in flood events.
After that, I use this non-linear similarity measure (ED) to compute the pairwise dependency in extreme precipitation event series. I incorporate the similarity measure within the framework of complex network theory to study the collective behavior of climate extremes. Under this architecture, the nodes are defined by the spatial grid points of the given spatio-temporal climate dataset. Each node is associated with a time series corresponding to the temporal evolution
of the climate observation at that grid point. Finally, the network links are functions of the pairwise statistical interdependence between the nodes. Various network measures, such as degree, betweenness centrality, clustering coefficient, etc., can be used to quantify the network’s topology. We apply the methodology mentioned above to study the spatio-temporal coherence pattern of extreme rainfall events in the United States and the Ganga River basin, which reveals its relation to various climate processes and the orography of the region.
The identification of precursors associated with the occurrence of extreme events in the near future is extremely important to prepare the masses for an upcoming disaster and mitigate the potential risks associated with such events. Under this motivation, I propose an in-data prediction recipe for predicting the data structures that typically occur prior to extreme events using the Echo state network, a type of Recurrent Neural Network which is a part of the reservoir
computing framework. However, unlike previous works that identify precursory structures in the same variable in which extreme events are manifested (active variable), I try to predict these structures by using data from another dynamic variable (passive variable) which does not show large excursions from the nominal condition but carries imprints of these extreme events. Furthermore, my results demonstrate that the quality of prediction depends on the magnitude
of events, i.e., the higher the magnitude of the extreme, the better is its predictability skill. I show quantitatively that this is because the input signals collectively form a more coherent pattern for an extreme event of higher magnitude, which enhances the efficiency of the machine to predict the forthcoming extreme events.
The process of number symbolization is assumed to be critically influenced by the acquisition of so-called verbal number skills (e.g., verbally reciting the number chain and naming Arabic numerals). For the acquisition of these verbal number skills, verbal and visuospatial skills are discussed as contributing factors. In this context, children’s verbal number skills have been found to be associated with their concurrent spatial language skills such as mastery of verbal descriptions of spatial position (e.g., in front of, behind). In a longitudinal study with three measurement times (T1, T2, T3) at an interval of about 6 months, we evaluated the predictive role of preschool children’s (mean age at T1: 3 years and 10 months) spatial language skills for the acquisition of verbal number skills. Children’s spatial language skills at T2 significantly predicted their verbal number skills at T3, when controlling for influences of important covariates such as vocabulary knowledge. In addition, further analyses replicated previous results indicating that children’s spatial language skills at T2 were associated with their verbal number skills at T2. Exploratory analyses further revealed that children’s verbal number skills at T1 predict their spatial language at T2. Results suggests that better spatial language skills at the age of 4 years facilitate the future acquisition of verbal number skills.
Numerical magnitude information is assumed to be spatially represented in the form of a mental number line defined with respect to a body-centred, egocentric frame of reference. In this context, spatial language skills such as mastery of verbal descriptions of spatial position (e.g., in front of, behind, to the right/left) have been proposed to be relevant for grasping spatial relations between numerical magnitudes on the mental number line. We examined 4- to 5-year-old’s spatial language skills in tasks that allow responses in egocentric and allocentric frames of reference, as well as their relative understanding of numerical magnitude (assessed by a number word comparison task). In addition, we evaluated influences of children’s absolute understanding of numerical magnitude assessed by their number word comprehension (montring different numbers using their fingers) and of their knowledge on numerical sequences (determining predecessors and successors as well as identifying missing dice patterns of a series). Results indicated that when considering responses that corresponded to the egocentric perspective, children’s spatial language was associated significantly with their relative numerical magnitude understanding, even after controlling for covariates, such as children’s SES, mental rotation skills, and also absolute magnitude understanding or knowledge on numerical sequences. This suggests that the use of egocentric reference frames in spatial language may facilitate spatial representation of numbers along a mental number line and thus seem important for preschoolers’ relative understanding of numerical magnitude.
The highly structured nature of the educational sector demands effective policy mechanisms close to the needs of the field. That is why evidence-based policy making, endorsed by the European Commission under Erasmus+ Key Action 3, aims to make an alignment between the domains of policy and practice. Against this background, this article addresses two issues: First, that there is a vertical gap in the translation of higher-level policies to local strategies and regulations. Second, that there is a horizontal gap between educational domains regarding the policy awareness of individual players. This was analyzed in quantitative and qualitative studies with domain experts from the fields of virtual mobility and teacher training. From our findings, we argue that the combination of both gaps puts the academic bridge from secondary to tertiary education at risk, including the associated knowledge proficiency levels. We discuss the role of digitalization in the academic bridge by asking the question: which value does the involved stakeholders expect from educational policies? As a theoretical basis, we rely on the model of value co-creation for and by stakeholders. We describe the used instruments along with the obtained results and proposed benefits. Moreover, we reflect on the methodology applied, and we finally derive recommendations for future academic bridge policies.
Extreme habitats often harbor specific communities that differ substantially from non-extreme habitats. In many cases, these communities are characterized by archaea, bacteria and protists, whereas the number of species of metazoa and higher plants is relatively low. In extremely acidic habitats, mostly prokaryotes and protists thrive, and only very few metazoa thrive, for example, rotifers. Since many studies have investigated the physiology and ecology of individual species, there is still a gap in research on direct, trophic interactions among extremophiles. To fill this gap, we experimentally studied the trophic interactions between a predatory protist (Actinophrys sol, Heliozoa) and its prey, the rotifers Elosa woralli and Cephalodella sp., the ciliate Urosomoida sp. and the mixotrophic protist Chlamydomonas acidophila (a green phytoflagellate, Chlorophyta). We found substantial predation pressure on all animal prey. High densities of Chlamydomonas acidophila reduced the predation impact on the rotifers by interfering with the feeding behaviour of A. sol. These trophic relations represent a natural case of intraguild predation, with Chlamydomonas acidophila being the common prey and the rotifers/ciliate and A. sol being the intraguild prey and predator, respectively. We further studied this intraguild predation along a resource gradient using Cephalodella sp. as the intraguild prey. The interactions among the three species led to an increase in relative rotifer abundance with increasing resource (Chlamydomonas) densities. By applying a series of laboratory experiments, we revealed the complexity of trophic interactions within a natural extremophilic community.
It is well-known that individuals with aphasia (IWA) have difficulties understanding sentences that involve non-adjacent dependencies, such as object relative clauses or passives (Caplan, Baker, & Dehaut, 1985; Caramazza & Zurif, 1976). A large body of research supports the view that IWA’s grammatical system is intact, and that comprehension difficulties in aphasia are caused by a processing deficit, such as a delay in lexical access and/or in syntactic structure building (e.g., Burkhardt, Piñango, & Wong, 2003; Caplan, Michaud, & Hufford, 2015; Caplan, Waters, DeDe, Michaud, & Reddy, 2007; Ferrill, Love, Walenski, & Shapiro, 2012; Hanne, Burchert, De Bleser, & Vasishth, 2015; Love, Swinney, Walenski, & Zurif, 2008). The main goal of this dissertation is to computationally investigate the processing sources of comprehension impairments in sentence processing in aphasia.
In this work, prominent theories of processing deficits coming from the aphasia literature are implemented within two cognitive models of sentence processing –the activation-based model (Lewis & Vasishth, 2005) and the direct-access model (McEl- ree, 2000)–. These models are two different expressions of the cue-based retrieval theory (Lewis, Vasishth, & Van Dyke, 2006), which posits that sentence processing is the result of a series of iterative retrievals from memory. These two models have been widely used to account for sentence processing in unimpaired populations in multiple languages and linguistic constructions, sometimes interchangeably (Parker, Shvarts- man, & Van Dyke, 2017). However, Nicenboim and Vasishth (2018) showed that when both models are implemented in the same framework and fitted to the same data, the models yield different results, because the models assume different data- generating processes. Specifically, the models hold different assumptions regarding the retrieval latencies. The second goal of this dissertation is to compare these two models of cue-based retrieval, using data from individuals with aphasia and control participants. We seek to answer the following question: Which retrieval mechanism is more likely to mediate sentence comprehension?
We model 4 subsets of existing data: Relative clauses in English and German; and control structures and pronoun resolution in German. The online data come from either self-paced listening experiments, or visual-world eye-tracking experiments. The offline data come from a complementary sentence-picture matching task performed at the end of the trial in both types of experiments. The two competing models of retrieval are implemented in the Bayesian framework, following Nicenboim and Vasishth (2018). In addition, we present a modified version of the direct-acess model that – we argue – is more suitable for individuals with aphasia.
This dissertation presents a systematic approach to implement and test verbally- stated theories of comprehension deficits in aphasia within cognitive models of sen- tence processing. The conclusions drawn from this work are that (a) the original direct-access model (as implemented here) cannot account for the full pattern of data from individuals with aphasia because it cannot account for slow misinterpretations; and (b) an activation-based model of retrieval can account for sentence comprehension deficits in individuals with aphasia by assuming a delay in syntactic structure building, and noise in the processing system. The overall pattern of results support an activation-based mechanism of memory retrieval, in which a combination of processing deficits, namely slow syntax and intermittent deficiencies, cause comprehension difficulties in individuals with aphasia.
The prevalence of obesity in the pediatric population has become a major public health issue. Indeed, the dramatic increase of this epidemic causes multiple and harmful consequences, Physical activity, particularly physical exercise, remains to be the cornerstone of interventions against childhood obesity. Given the conflicting findings with reference to the relevant literature addressing the effects of exercise on adiposity and physical fitness outcomes in obese children and adolescents, the effect of duration-matched concurrent training (CT) [50% resistance (RT) and 50% high-intensity-interval-training (HIIT)] on body composition and physical fitness in obese youth remains to be elucidated. Thus, the purpose of this study was to examine the effects of 9-weeks of CT compared to RT or HIIT alone, on body composition and selected physical fitness components in healthy sedentary obese youth. Out of 73 participants, only 37; [14 males and 23 females; age 13.4 ± 0.9 years; body-mass-index (BMI): 31.2 ± 4.8 kg·m-2] were eligible and randomized into three groups: HIIT (n = 12): 3-4 sets×12 runs at 80–110% peak velocity, with 10-s passive recovery between bouts; RT (n = 12): 6 exercises; 3–4 sets × 10 repetition maximum (RM) and CT (n = 13): 50% serial completion of RT and HIIT. CT promoted significant greater gains compared to HIIT and RT on body composition (p < 0.01, d = large), 6-min-walking test distance (6 MWT-distance) and on 6 MWT-VO2max (p < 0.03, d = large). In addition, CT showed substantially greater improvements than HIIT in the medicine ball throw test (20.2 vs. 13.6%, p < 0.04, d = large). On the other hand, RT exhibited significantly greater gains in relative hand grip strength (p < 0.03, d = large) and CMJ (p < 0.01, d = large) than HIIT and CT. CT promoted greater benefits for fat, body mass loss and cardiorespiratory fitness than HIIT or RT modalities. This study provides important information for practitioners and therapists on the application of effective exercise regimes with obese youth to induce significant and beneficial body composition changes. The applied CT program and the respective programming parameters in terms of exercise intensity and volume can be used by practitioners as an effective exercise treatment to fight the pandemic overweight and obesity in youth.
Localisation of deformation is a ubiquitous feature in continental rift dynamics and observed across drastically different time and length scales. This thesis comprises one experimental and two numerical modelling studies investigating strain localisation in (1) a ductile shear zone induced by a material heterogeneity and (2) in an active continental rift setting. The studies are related by the fact that the weakening mechanisms on the crystallographic and grain size scale enable bulk rock weakening, which fundamentally enables the formation of shear zones, continental rifts and hence plate tectonics. Aiming to investigate the controlling mechanisms on initiation and evolution of a shear zone, the torsion experiments of the experimental study were conducted in a Patterson type apparatus with strong Carrara marble cylinders with a weak, planar Solnhofen limestone inclusion. Using state-of-the-art numerical modelling software, the torsion experiments were simulated to answer questions regarding localisation procedure like stress distribution or the impact of rheological weakening. 2D numerical models were also employed to integrate geophysical and geological data to explain characteristic tectonic evolution of the Southern and Central Kenya Rift. Key elements of the numerical tools are a randomized initial strain distribution and the usage of strain softening. During the torsion experiments, deformation begins to localise at the limestone inclusion tips in a process zone, which propagates into the marble matrix with increasing deformation until a ductile shear zone is established. Minor indicators for coexisting brittle deformation are found close to the inclusion tip and presumed to slightly facilitate strain localisation besides the dominant ductile deformation processes. The 2D numerical model of the torsion experiment successfully predicts local stress concentration and strain rate amplification ahead of the inclusion in first order agreement with the experimental results. A simple linear parametrization of strain weaking enables high accuracy reproduction of phenomenological aspects of the observed weakening. The torsion experiments suggest that loading conditions do not affect strain localisation during high temperature deformation of multiphase material with high viscosity contrasts. A numerical simulation can provide a way of analysing the process zone evolution virtually and extend the examinable frame. Furthermore, the nested structure and anastomosing shape of an ultramylonite band was mimicked with an additional second softening step. Rheological weakening is necessary to establish a shear zone in a strong matrix around a weak inclusion and for ultramylonite formation.
Such strain weakening laws are also incorporated into the numerical models of the
Southern and Central Kenya Rift that capture the characteristic tectonic evolution. A three-stage early rift evolution is suggested that starts with (1) the accommodation of strain by a single border fault and flexure of the hanging-wall crust, after which (2) faulting in the hanging-wall and the basin centre increases before (3) the early-stage asymmetry is lost and basinward localisation of deformation occurs. Along-strike variability of rifts can be produced by modifying the initial random noise distribution. In summary, the three studies address selected aspects of the broad range of mechanisms and processes that fundamentally enable the deformation of rock and govern the localisation patterns across the scales. In addition to the aforementioned results, the first and second manuscripts combined, demonstrate a procedure to find new or improve on existing numerical formulations for specific rheologies and their dynamic weakening. These formulations are essential in addressing rock deformation from the grain to the global scale. As within the third study of this thesis, where geodynamic controls on the evolution of a rift were examined and acquired by the integration of geological and geophysical data into a numerical model.
This thesis is analyzing multiple coordination challenges which arise with the digital transformation of public administration in federal systems, illustrated by four case studies in Germany. I make various observations within a multi-level system and provide an in-depth analysis. Theoretical explanations from both federalism research and neo-institutionalism are utilized to explain the findings of the empirical driven work. The four articles evince a holistic picture of the German case and elucidate its role as a digital government laggard. Their foci range from macro, over meso to micro level of public administration, differentiating between the governance and the tool dimension of digital government.
The first article shows how multi-level negotiations lead to expensive but eventually satisfying solutions for the involved actors, creating a subtle balance between centralization and decentralization. The second article identifies legal, technical, and organizational barriers for cross-organizational service provision, highlighting the importance of inter-organizational and inter-disciplinary exchange and both a common language and trust. Institutional change and its effects on the micro level, on citizens and the employees in local one-stop shops, mark the focus of the third article, bridging the gap between reforms and the administrative reality on the local level. The fourth article looks at the citizens’ perspective on digital government reforms, their expectations, use and satisfaction. In this vein, this thesis provides a detailed account of the importance of understanding the digital divide and therefore the necessity of reaching out to different recipients of digital government reforms. I draw conclusions from the factors identified as causes for Germany’s shortcomings for other federal systems where feasible and derive reform potential therefrom. This allows to gain a new perspective on digital government and its coordination challenges in federal contexts.
Core-shell upconversion nanoparticles - investigation of dopant intermixing and surface modification
(2022)
Frequency upconversion nanoparticles (UCNPs) are inorganic nanocrystals capable to up-convert incident photons of the near-infrared electromagnetic spectrum (NIR) into higher energy photons. These photons are re-emitted in the range of the visible (Vis) and even ultraviolet (UV) light. The frequency upconversion process (UC) is realized with nanocrystals doped with trivalent lanthanoid ions (Ln(III)). The Ln(III) ions provide the electronic (excited) states forming a ladder-like electronic structure for the Ln(III) electrons in the nanocrystals. The absorption of at least two low energy photons by the nanoparticle and the subsequent energy transfer to one Ln(III) ion leads to the promotion of one Ln(III) electron into higher excited electronic states. One high energy photon will be emitted during the radiative relaxation of the electron in the excited state back into the electronic ground state of the Ln(III) ion. The excited state electron is the result of the previous absorption of at least two low energy photons.
The UC process is very interesting in the biological/medical context. Biological samples (like organic tissue, blood, urine, and stool) absorb high-energy photons (UV and blue light) more strongly than low-energy photons (red and NIR light). Thanks to a naturally occurring optical window, NIR light can penetrate deeper than UV light into biological samples. Hence, UCNPs in bio-samples can be excited by NIR light. This possibility opens a pathway for in vitro as well as in vivo applications, like optical imaging by cell labeling or staining of specific organic tissue. Furthermore, early detection and diagnosis of diseases by predictive and diagnostic biomarkers can be realized with bio-recognition elements being labeled to the UCNPs. Additionally, "theranostic" becomes possible, in which the identification and the treatment of a disease are tackled simultaneously.
For this to succeed, certain parameters for the UCNPs must be met: high upconversion efficiency, high photoluminescence quantum yield, dispersibility, and dispersion stability in aqueous media, as well as availability of functional groups to introduce fast and easy bio-recognition elements. The UCNPs used in this work were prepared with a solvothermal decomposition synthesis yielding in particles with NaYF4 or NaGdF4 as host lattice. They have been doped with the Ln(III) ions Yb3+ and Er3+, which is only one possible upconversion pair. Their upconversion efficiency and photoluminescence quantum yield were improved by adding a passivating shell to reduce surface quenching.
However, the brightness of core-shell UCNPs stays behind the expectations compared to their bulk material (being at least μm-sized particles). The core-shell structures are not clearly separated from each other, which is a topic in literature. Instead, there is a transition layer between the core and the shell structure, which relates to the migration of the dopants within the host lattice during the synthesis. The ion migration has been examined by time-resolved laser spectroscopy and the interlanthanoid resonance energy transfer (LRET) in the two different host lattices from above. The results are
presented in two publications, which dealt with core-shell-shell structured nanoparticles. The core is doped with the LRET-acceptor (either Nd3+ or Pr3+). The intermediate shell serves as an insulation shell of pure host lattice material, whose shell thickness has been varied within one set of samples having the same composition, so that the spatial separation of LRET-acceptor and -donor changes. The outer shell with the same host lattice is doped with the LRET-donor (Eu3+). The effect of the increasing insulation shell thickness is significant, although the LRET cannot be suppressed completely.
Next to the Ln(III) migration within a host lattice, various phase transfer reactions were investigated in order to subsequently perform surface modifications for bioapplications. One result out of this research has been published using a promising ligand, that equips the UCNP with bio-modifiable groups and has good potential for bio-medical applications. This particular ligand mimics natural occurring mechanisms of mussel protein adhesion and of blood coagulation, which is why the UCNPs are encapsulated very effectively. At the same time, bio-functional groups are introduced. In a proof-of-concept, the encapsulated UCNP has been coupled successfully with a dye (which is representative for a biomarker) and the system’s photoluminescence properties have been investigated.
The motivation for this work was the question of reliability and robustness of seismic tomography. The problem is that many earth models exist which can describe the underlying ground motion records equally well. Most algorithms for reconstructing earth models provide a solution, but rarely quantify their variability. If there is no way to verify the imaged structures, an interpretation is hardly reliable. The initial idea was to explore the space of equivalent earth models using Bayesian inference. However, it quickly became apparent that the rigorous quantification of tomographic uncertainties could not be accomplished within the scope of a dissertation.
In order to maintain the fundamental concept of statistical inference, less complex problems from the geosciences are treated instead. This dissertation aims to anchor Bayesian inference more deeply in the geosciences and to transfer knowledge from applied mathematics. The underlying idea is to use well-known methods and techniques from statistics to quantify the uncertainties of inverse problems in the geosciences. This work is divided into three parts:
Part I introduces the necessary mathematics and should be understood as a kind of toolbox. With a physical application in mind, this section provides a compact summary of all methods and techniques used. The introduction of Bayesian inference makes the beginning. Then, as a special case, the focus is on regression with Gaussian processes under linear transformations. The chapters on the derivation of covariance functions and the approximation of non-linearities are discussed in more detail.
Part II presents two proof of concept studies in the field of seismology. The aim is to present the conceptual application of the introduced methods and techniques with moderate complexity. The example about traveltime tomography applies the approximation of non-linear relationships. The derivation of a covariance function using the wave equation is shown in the example of a damped vibrating string. With these two synthetic applications, a consistent concept for the quantification of modeling uncertainties has been developed.
Part III presents the reconstruction of the Earth's archeomagnetic field. This application uses the whole toolbox presented in Part I and is correspondingly complex. The modeling of the past 1000 years is based on real data and reliably quantifies the spatial modeling uncertainties. The statistical model presented is widely used and is under active development.
The three applications mentioned are intentionally kept flexible to allow transferability to similar problems. The entire work focuses on the non-uniqueness of inverse problems in the geosciences. It is intended to be of relevance to those interested in the concepts of Bayesian inference.
The geomagnetic main field is vital for live on Earth, as it shields our habitat against the solar wind and cosmic rays. It is generated by the geodynamo in the Earth’s outer core and has a rich dynamic on various timescales. Global models of the field are used to study the interaction of the field and incoming charged particles, but also to infer core dynamics and to feed numerical simulations of the geodynamo. Modern satellite missions, such as the SWARM or the CHAMP mission, support high resolution reconstructions of the global field. From the 19 th century on, a global network of magnetic observatories has been established. It is growing ever since and global models can be constructed from the data it provides. Geomagnetic field models that extend further back in time rely on indirect observations of the field, i.e. thermoremanent records such as burnt clay or volcanic rocks and sediment records from lakes and seas. These indirect records come with (partially very large) uncertainties, introduced by the complex measurement methods and the dating procedure.
Focusing on thermoremanent records only, the aim of this thesis is the development of a new modeling strategy for the global geomagnetic field during the Holocene, which takes the uncertainties into account and produces realistic estimates of the reliability of the model. This aim is approached by first considering snapshot models, in order to address the irregular spatial distribution of the records and the non-linear relation of the indirect observations to the field itself. In a Bayesian setting, a modeling algorithm based on Gaussian process regression is developed and applied to binned data. The modeling algorithm is then extended to the temporal domain and expanded to incorporate dating uncertainties. Finally, the algorithm is sequentialized to deal with numerical challenges arising from the size of the Holocene dataset.
The central result of this thesis, including all of the aspects mentioned, is a new global geomagnetic field model. It covers the whole Holocene, back until 12000 BCE, and we call it ArchKalmag14k. When considering the uncertainties that are produced together with the model, it is evident that before 6000 BCE the thermoremanent database is not sufficient to support global models. For times more recent, ArchKalmag14k can be used to analyze features of the field under consideration of posterior uncertainties. The algorithm for generating ArchKalmag14k can be applied to different datasets and is provided to the community as an open source python package.
The complex hierarchical structure of bone undergoes a lifelong remodeling process, where it adapts to mechanical needs. Hereby, bone resorption by osteoclasts and bone formation by osteoblasts have to be balanced to sustain a healthy and stable organ. Osteocytes orchestrate this interplay by sensing mechanical strains and translating them into biochemical signals. The osteocytes are located in lacunae and are connected to one another and other bone cells via cell processes through small channels, the canaliculi. Lacunae and canaliculi form a network (LCN) of extracellular spaces that is able to transport ions and enables cell-to-cell communication. Osteocytes might also contribute to mineral homeostasis by direct interactions with the surrounding matrix. If the LCN is acting as a transport system, this should be reflected in the mineralization pattern. The central hypothesis of this thesis is that osteocytes are actively changing their material environment. Characterization methods of material science are used to achieve the aim of detecting traces of this interaction between osteocytes and the extracellular matrix. First, healthy murine bones were characterized. The properties analyzed were then compared with three murine model systems: 1) a loading model, where a bone of the mouse was loaded during its life time; 2) a healing model, where a bone of the mouse was cut to induce a healing response; and 3) a disease model, where the Fbn1 gene is dysfunctional causing defects in the formation of the extracellular tissue.
The measurement strategy included routines that make it possible to analyze the organization of the LCN and the material components (i.e., the organic collagen matrix and the mineral particles) in the same bone volumes and compare the spatial distribution of different data sets. The three-dimensional network architecture of the LCN is visualized by confocal laser scanning microscopy (CLSM) after rhodamine staining and is then subsequently quantified. The calcium content is determined via quantitative backscattered electron imaging (qBEI), while small- and wide-angle X-ray scattering (SAXS and WAXS) are employed to determine the thickness and length of local mineral particles.
First, tibiae cortices of healthy mice were characterized to investigate how changes in LCN architecture can be attributed to interactions of osteocytes with the surrounding bone matrix. The tibial mid-shaft cross-sections showed two main regions, consisting of a band with unordered LCN surrounded by a region with ordered LCN. The unordered region is a remnant of early bone formation and exhibited short and thin mineral particles. The surrounding, more aligned bone showed ordered and dense LCN as well as thicker and longer mineral particles. The calcium content was unchanged between the two regions.
In the mouse loading model, the left tibia underwent two weeks of mechanical stimulation, which results in increased bone formation and decreased resorption in skeletally mature mice. Here the specific research question addressed was how do bone material characteristics change at (re)modeling sites? The new bone formed in response to mechanical stimulation showed similar properties in terms of the mineral particles, like the ordered calcium region but lower calcium content compared to the right, non-loaded control bone of the same mice. There was a clear, recognizable border between mature and newly formed bone. Nevertheless, some canaliculi went through this border connecting the LCN of mature and newly formed bone.
Additionally, the question should be answered whether the LCN topology and the bone matrix material properties adapt to loading. Although, mechanically stimulated bones did not show differences in calcium content compared to controls, different correlations were found between the local LCN density and the local Ca content depending on whether the bone was loaded or not. These results suggest that the LCN may serve as a mineral reservoir.
For the healing model, the femurs of mice underwent an osteotomy, stabilized with an external fixator and were allowed to heal for 21 days. Thus, the spatial variations in the LCN topology with mineral properties within different tissue types and their interfaces, namely calcified cartilage, bony callus and cortex, could be simultaneously visualized and compared in this model. All tissue types showed structural differences across multiple length scales. Calcium content increased and became more homogeneous from calcified cartilage to bony callus to lamellar cortical bone. The degree of LCN organization increased as well, while the lacunae became smaller, as did the lacunar density between these different tissue types that make up the callus. In the calcified cartilage, the mineral particles were short and thin. The newly formed callus exhibited thicker mineral particles, which still had a low degree of orientation. While most of the callus had a woven-like structure, it also served as a scaffold for more lamellar tissue at the edges. The lamelar bone callus showed thinner mineral particles, but a higher degree of alignment in both, mineral particles and the LCN. The cortex showed the highest values for mineral length, thickness and degree of orientation. At the same time, the lacunae number density was 34% lower and the lacunar volume 40% smaller compared to bony callus. The transition zone between cortical and callus regions showed a continuous convergence of bone mineral properties and lacunae shape. Although only a few canaliculi connected callus and the cortical region, this indicates that communication between osteocytes of both tissues should be possible. The presented correlations between LCN architecture and mineral properties across tissue types may suggest that osteocytes have an active role in mineralization processes of healing.
A mouse model for the disease marfan syndrome, which includes a genetic defect in the fibrillin-1 gene, was investigated. In humans, Marfan syndrome is characterized by a range of clinical symptoms such as long bone overgrowth, loose joints, reduced bone mineral density, compromised bone microarchitecture, and increased fracture rates. Thus, fibrillin-1 seems to play a role in the skeletal homeostasis. Therefore, the present work studied how marfan syndrome alters LCN architecture and the surrounding bone matrix. The mice with marfan syndrome showed longer tibiae than their healthy littermates from an age of seven weeks onwards. In contrast, the cortical development appeared retarded, which was observed across all measured characteristics, i. e. lower endocortical bone formation, looser and less organized lacuno-canalicular network, less collagen orientation, thinner and shorter mineral particles.
In each of the three model systems, this study found that changes in the LCN architecture spatially correlated with bone matrix material parameters. While not knowing the exact mechanism, these results provide indications that osteocytes can actively manipulate a mineral reservoir located around the canaliculi to make a quickly accessible contribution to mineral homeostasis. However, this interaction is most likely not one-sided, but could be understood as an interplay between osteocytes and extra-cellular matrix, since the bone matrix contains biochemical signaling molecules (e.g. non-collagenous proteins) that can change osteocyte behavior. Bone (re)modeling can therefore not only be understood as a method for removing defects or adapting to external mechanical stimuli, but also for increasing the efficiency of possible osteocyte-mineral interactions during bone homeostasis. With these findings, it seems reasonable to consider osteocytes as a target for drug development related to bone diseases that cause changes in bone composition and mechanical properties. It will most likely require the combined effort of materials scientists, cell biologists, and molecular biologists to gain a deeper understanding of how bone cells respond to their material environment.
Cosmic rays (CRs) are a ubiquitous and an important component of astrophysical environments such as the interstellar medium (ISM) and intracluster medium (ICM). Their plasma physical interactions with electromagnetic fields strongly influence their transport properties. Effective models which incorporate the microphysics of CR transport are needed to study the effects of CRs on their surrounding macrophysical media. Developing such models is challenging because of the conceptional, length-scale, and time-scale separation between the microscales of plasma physics and the macroscales of the environment. Hydrodynamical theories of CR transport achieve this by capturing the evolution of CR population in terms of statistical moments. In the well-established one-moment hydrodynamical model for CR transport, the dynamics of the entire CR population are described by a single statistical quantity such as the commonly used CR energy density. In this work, I develop a new hydrodynamical two-moment theory for CR transport that expands the well-established hydrodynamical model by including the CR energy flux as a second independent hydrodynamical quantity. I detail how this model accounts for the interaction between CRs and gyroresonant Alfvén waves. The small-scale magnetic fields associated with these Alfvén waves scatter CRs which fundamentally alters CR transport along large-scale magnetic field lines. This leads to the effects of CR streaming and diffusion which are both captured within the presented hydrodynamical theory. I use an Eddington-like approximation to close the hydrodynamical equations and investigate the accuracy of this closure-relation by comparing it to high-order approximations of CR transport. In addition, I develop a finite-volume scheme for the new hydrodynamical model and adapt it to the moving-mesh code Arepo. This scheme is applied using a simulation of a CR-driven galactic wind. I investigate how CRs launch the wind and perform a statistical analysis of CR transport properties inside the simulated circumgalactic medium (CGM). I show that the new hydrodynamical model can be used to explain the morphological appearance of a particular type of radio filamentary structures found inside the central molecular zone (CMZ). I argue that these harp-like features are synchrotron-radiating CRs which are injected into braided magnetic field lines by a point-like source such as a stellar wind of a massive star or a pulsar. Lastly, I present the finite-volume code Blinc that uses adaptive mesh refinement (AMR) techniques to perform simulations of radiation and magnetohydrodynamics (MHD). The mesh of Blinc is block-structured and represented in computer memory using a graph-based approach. I describe the implementation of the mesh graph and how a diffusion process is employed to achieve load balancing in parallel computing environments. Various test problems are used to verify the accuracy and robustness of the employed numerical algorithms.
High-throughput proteomics approaches have resulted in large-scale protein–protein interaction (PPI) networks that have been employed for the prediction of protein complexes. However, PPI networks contain false-positive as well as false-negative PPIs that affect the protein complex prediction algorithms. To address this issue, here we propose an algorithm called CUBCO+ that: (1) employs GO semantic similarity to retain only biologically relevant interactions with a high similarity score, (2) based on link prediction approaches, scores the false-negative edges, and (3) incorporates the resulting scores to predict protein complexes. Through comprehensive analyses with PPIs from Escherichia coli, Saccharomyces cerevisiae, and Homo sapiens, we show that CUBCO+ performs as well as the approaches that predict protein complexes based on recently introduced graph partitions into biclique spanned subgraphs and outperforms the other state-of-the-art approaches. Moreover, we illustrate that in combination with GO semantic similarity, CUBCO+ enables us to predict more accurate protein complexes in 36% of the cases in comparison to CUBCO as its predecessor.
Teaching and learning as well as administrative processes are still experiencing intensive changes with the rise of artificial intelligence (AI) technologies and its diverse application opportunities in the context of higher education. Therewith, the scientific interest in the topic in general, but also specific focal points rose as well. However, there is no structured overview on AI in teaching and administration processes in higher education institutions that allows to identify major research topics and trends, and concretizing peculiarities and develops recommendations for further action. To overcome this gap, this study seeks to systematize the current scientific discourse on AI in teaching and administration in higher education institutions. This study identified an (1) imbalance in research on AI in educational and administrative contexts, (2) an imbalance in disciplines and lack of interdisciplinary research, (3) inequalities in cross-national research activities, as well as (4) neglected research topics and paths. In this way, a comparative analysis between AI usage in administration and teaching and learning processes, a systematization of the state of research, an identification of research gaps as well as further research path on AI in higher education institutions are contributed to research.
This dataset comprises tree inventories and damage assessments performed in Namibia's semi-arid Zambezi Region. Data were sampled in savannas and savanna woodlands along steep gradients of elephant population densities to capture the effects of those (and other) disturbances on individual-level and stand-level aboveground woody biomass (AGB). The dataset contains raw data on dendrometric measures and processed data on specific wood density (SWD), woody aboveground biomass, and biomass losses through disturbance impacts. Allometric proxies (height, canopy diameters, and in adult trees also stem circumferences) were recorded for n = 6,179 tree and shrub individuals. Wood samples were taken for each encountered species to measure specific wood density.
These measurements have been used to estimate woody aboveground biomass via established allometric models, advanced through our improved methodologies and workflows that accounted for tree and shrub architecture shaped by disturbance impacts. To this end, we performed a detailed damage assessment on each woody individual in the field. In addition to estimations of standing biomass, our new method also delivered data on biomass losses to different disturbance agents (elephants, fire, and others) on the level of plant individuals and stands.
The data presented here have been used within a study published with Ecological Indicators (Kindermann et al., 2022) to evaluate the benefits of our improved methodology in comparison to a standard reference method of aboveground biomass estimations. Additionally, it has been employed in a study on carbon storage and sequestration in vegetation and soils (Sandhage-Hofmann et al., 2021).
The raw data of dendrometric measurements can be subjected to other available allometric models for biomass estimation. The processed data can be used to analyze disturbance impacts on woody aboveground biomass, or for regional carbon storage estimates. The data on species-specific wood density can be used for application to other dendrometric datasets to (re-) estimate biomass through allometric models requiring wood density. It can further be used for plant functional trait analyses.
Cutting-edge hyperscanning methods led to a paradigm shift in social neuroscience. It allowed researchers to measure dynamic mutual alignment of neural processes between two or more individuals in naturalistic contexts. The ever-growing interest in hyperscanning research calls for the development of transparent and validated data analysis methods to further advance the field. We have developed and tested a dual electroencephalography (EEG) analysis pipeline, namely DEEP. Following the preprocessing of the data, DEEP allows users to calculate Phase Locking Values (PLVs) and cross-frequency PLVs as indices of inter-brain phase alignment of dyads as well as time-frequency responses and EEG power for each participant. The pipeline also includes scripts to control for spurious correlations. Our goal is to contribute to open and reproducible science practices by making DEEP publicly available together with an example mother-infant EEG hyperscanning dataset.
Privacy regulations and the physical distribution of heterogeneous data are often primary concerns for the development of deep learning models in a medical context. This paper evaluates the feasibility of differentially private federated learning for chest X-ray classification as a defense against data privacy attacks. To the best of our knowledge, we are the first to directly compare the impact of differentially private training on two different neural network architectures, DenseNet121 and ResNet50. Extending the federated learning environments previously analyzed in terms of privacy, we simulated a heterogeneous and imbalanced federated setting by distributing images from the public CheXpert and Mendeley chest X-ray datasets unevenly among 36 clients. Both non-private baseline models achieved an area under the receiver operating characteristic curve (AUC) of 0.940.94 on the binary classification task of detecting the presence of a medical finding. We demonstrate that both model architectures are vulnerable to privacy violation by applying image reconstruction attacks to local model updates from individual clients. The attack was particularly successful during later training stages. To mitigate the risk of a privacy breach, we integrated Rényi differential privacy with a Gaussian noise mechanism into local model training. We evaluate model performance and attack vulnerability for privacy budgets ε∈{1,3,6,10}�∈{1,3,6,10}. The DenseNet121 achieved the best utility-privacy trade-off with an AUC of 0.940.94 for ε=6�=6. Model performance deteriorated slightly for individual clients compared to the non-private baseline. The ResNet50 only reached an AUC of 0.760.76 in the same privacy setting. Its performance was inferior to that of the DenseNet121 for all considered privacy constraints, suggesting that the DenseNet121 architecture is more robust to differentially private training.
Wheat alpha-amylase/trypsin inhibitors remain a subject of interest considering the latest findings showing their implication in wheat-related non-celiac sensitivity (NCWS). Understanding their functions in such a disorder is still unclear and for further study, the need for pure ATI molecules is one of the limiting problems. In this work, a simplified approach based on the successive fractionation of ATI extracts by reverse phase and ion exchange chromatography was developed. ATIs were first extracted from wheat flour using a combination of Tris buffer and chloroform/methanol methods. The separation of the extracts on a C18 column generated two main fractions of interest F1 and F2. The response surface methodology with the Doehlert design allowed optimizing the operating parameters of the strong anion exchange chromatography. Finally, the seven major wheat ATIs namely P01083, P17314, P16850, P01085, P16851, P16159, and P83207 were recovered with purity levels (according to the targeted LC-MS/MS analysis) of 98.2 ± 0.7; 98.1 ± 0.8; 97.9 ± 0.5; 95.1 ± 0.8; 98.3 ± 0.4; 96.9 ± 0.5, and 96.2 ± 0.4%, respectively. MALDI-TOF-MS analysis revealed single peaks in each of the pure fractions and the mass analysis yielded deviations of 0.4, 1.9, 0.1, 0.2, 0.2, 0.9, and 0.1% between the theoretical and the determined masses of P01083, P17314, P16850, P01085, P16851, P16159, and P83207, respectively. Overall, the study allowed establishing an efficient purification process of the most important wheat ATIs. This paves the way for further in-depth investigation of the ATIs to gain more knowledge related to their involvement in NCWS disease and to allow the absolute quantification in wheat samples.
These days design thinking is no longer a “new approach”. Among practitioners, as well as academics, interest in the topic has gathered pace over the last two decades. However, opinions are divided over the longevity of the phenomenon: whether design thinking is merely “old wine in new bottles,” a passing trend, or still evolving as it is being spread to an increasing number of organizations and industries. Despite its growing relevance and the diffusion of design thinking, knowledge on the actual status quo in organizations remains scarce. With a new study, the research team of Prof. Uebernickel and Stefanie Gerken investigates temporal developments and changes in design thinking practices in organizations over the past six years comparing the results of the 2015 “Parts without a whole” study with current practices and future developments. Companies of all sizes and from different parts of the world participated in the survey. The findings from qualitative interviews with experts, i.e., people who have years of knowledge with design thinking, were cross-checked with the results from an exploratory analysis of the survey data. This analysis uncovers significant variances and similarities in how design thinking is interpreted and applied in businesses.
Stellar interferometry is the only method in observational astronomy for obtaining the highest resolution images of astronomical targets. This method is based on combining light from two or more separate telescopes to obtain the complex visibility that contains information about the brightness distribution of an astronomical source. The applications of stellar interferometry have made significant contributions in the exciting research areas of astronomy and astrophysics, including the precise measurement of stellar diameters, imaging of stellar surfaces, observations of circumstellar disks around young stellar objects, predictions of Einstein's General relativity at the galactic center, and the direct search for exoplanets to name a few. One important related technique is aperture masking interferometry, pioneered in the 1960s, which uses a mask with holes at the re-imaged pupil of the telescope, where the light from the holes is combined using the principle of stellar interferometry. While this can increase the resolution, it comes with a disadvantage. Due to the finite size of the holes, the majority of the starlight (typically > 80 %) is lost at the mask, thus limiting the signal-to-noise ratio (SNR) of the output images. This restriction of aperture masking only to the bright targets can be avoided using pupil remapping interferometry - a technique combining aperture masking interferometry and advances in photonic technologies using single-mode fibers. Due to the inherent spatial filtering properties, the single-mode fibers can be placed at the focal plane of the re-imaged pupil, allowing the utilization of the whole pupil of the telescope to produce a high-dynamic range along with high-resolution images. Thus, pupil remapping interferometry is one of the most promising application areas in the emerging field of astrophotonics.
At the heart of an interferometric facility, a beam combiner exists whose primary function is to combine light to obtain high-contrast fringes. A beam combiner can be as simple as a beam splitter or an anamorphic lens to combine light from 2 apertures (or telescopes) or as complex as a cascade of beam splitters and lenses to combine light for > 2 apertures. However, with the field of astrophotonics, interferometric facilities across the globe are increasingly employing some form of photonics technologies by using single-mode fibers or integrated optics (IO) chips as an efficient way to combine light from several apertures. The state-of-the-art instrument - GRAVITY at the very large telescope interferometer (VLTI) facility uses an IO-based beam combiner device reaching visibilities accuracy of better than < 0.25 %, which is roughly 50× as precise as a few decades back.
Therefore, in the context of IO-based components for applications in stellar interferometry, this Thesis describes the work towards the development of a 3-dimensional (3-D) IO device - a monolithic astrophotonics component containing both the pupil remappers and a discrete beam combiner (DBC). In this work, the pupil remappers are 3-D single-mode waveguides in a glass substrate collecting light from the re-imaged pupil of the telescope and feeding the light to a DBC, where the combination takes place. The DBC is a lattice of 3-D single-mode waveguides, which interact through evanescent coupling. By observing the output power of single-mode waveguides of the DBC, the visibilities are retrieved by using a calibrated transfer matrix ({U}) of the device.
The feasibility of the DBC in retrieving the visibilities theoretically and experimentally had already been studied in the literature but was only limited to laboratory tests with monochromatic light sources. Thus, a part of this work extends these studies by investigating the response of a 4-input DBC to a broad-band light source. Hence, the objectives of this Thesis are the following: 1) Design an IO device for broad-band light operation such that accurate and precise visibilities could be retrieved experimentally at astronomical H-band (1.5-1.65 μm), and 2) Validation of the DBC as a possible beam combination scheme for future interferometric facilities through on-sky testing at the William Herschel Telescope (WHT).
This work consisted of designing three different 3-D IO devices. One of the popular methods for fabricating 3-D photonic components in a glass substrate is ultra-fast laser inscription (ULI). Thus, manufacturing of the designed devices was outsourced to Politecnico di Milano as part of an iterative fabrication process using their state-of-the-art ULI facility. The devices were then characterized using a 2-beam Michelson interferometric setup obtaining both the monochromatic and polychromatic visibilities. The retrieved visibilities for all devices were in good agreement as predicted by the simulation results of a DBC, which confirms both the repeatability of the ULI process and the stability of the Michelson setup, thus fulfilling the first objective.
The best-performing device was then selected for the pupil-remapping of the WHT using a different optical setup consisting of a deformable mirror and a microlens array. The device successfully collected stellar photons from Vega and Altair. The visibilities were retrieved using a previously calibrated {U} but showed significant deviations from the expected results. Based on the analysis of comparable simulations, it was found that such deviations were primarily caused by the limited SNR of the stellar observations, thus constituting a first step towards the fulfillment of the second objective.
Fiber-based microfluidics has undergone many innovative developments in recent years, with exciting examples of portable, cost-effective and easy-to-use detection systems already being used in diagnostic and analytical applications. In water samples, Legionella are a serious risk as human pathogens. Infection occurs through inhalation of aerosols containing Legionella cells and can cause severe pneumonia and may even be fatal. In case of Legionella contamination of water-bearing systems or Legionella infection, it is essential to find the source of the contamination as quickly as possible to prevent further infections. In drinking, industrial and wastewater monitoring, the culture-based method is still the most commonly used technique to detect Legionella contamination. In order to improve the laboratory-dependent determination, the long analysis times of 10-14 days as well as the inaccuracy of the measured values in colony forming units (CFU), new innovative ideas are needed. In all areas of application, for example in public, commercial or private facilities, rapid and precise analysis is required, ideally on site.
In this PhD thesis, all necessary single steps for a rapid DNA-based detection of Legionella were developed and characterized on a fiber-based miniaturized platform. In the first step, a fast, simple and device-independent chemical lysis of the bacteria and extraction of genomic DNA was established. Subsequently, different materials were investigated with respect to their non-specific DNA retention. Glass fiber filters proved to be particularly suitable, as they allow recovery of the DNA sample from the fiber material in combination with dedicated buffers and exhibit low autofluorescence, which was important for fluorescence-based readout.
A fiber-based electrophoresis unit was developed to migrate different oligonucleotides within a fiber matrix by application of an electric field. A particular advantage over lateral flow assays is the targeted movement, even after the fiber is saturated with liquid. For this purpose, the entire process of fiber selection, fiber chip patterning, combination with printed electrodes, and testing of retention and migration of different DNA samples (single-stranded, double-stranded and genomic DNA) was performed. DNA could be pulled across the fiber chip in an electric field of 24 V/cm within 5 minutes, remained intact and could be used for subsequent detection assays e.g., polymerase chain reaction (PCR) or fluorescence in situ hybridization (FISH). Fiber electrophoresis could also be used to separate DNA from other components e.g., proteins or cell lysates or to pull DNA through multiple layers of the glass microfiber. In this way, different fragments experienced a moderate, size-dependent separation. Furthermore, this arrangement offers the possibility that different detection reactions could take place in different layers at a later time. Electric current and potential measurements were collected to investigate the local distribution of the sample during migration. While an increase in current signal at high concentrations indicated the presence of DNA samples, initial experiments with methylene blue stained DNA showed a temporal sequence of signals, indicating sample migration along the chip.
For the specific detection of a Legionella DNA, a FISH-based detection with a molecular beacon probe was tested on the glass microfiber. A specific region within the 16S rRNA gene of Legionella spp. served as a target. For this detection, suitable reaction conditions and a readout unit had to be set up first. Subsequently, the sensitivity of the probe was tested with the reverse complementary target sequence and the specificity with several DNA fragments that differed from the target sequence. Compared to other DNA sequences of similar length also found in Legionella pneumophila, only the target DNA was specifically detected on the glass microfiber. If a single base exchange is present or if two bases are changed, the probe can no longer distinguish between the DNA targets and non-targets. An analysis with this specificity can be achieved with other methods such as melting point determination, as was also briefly indicated here. The molecular beacon probe could be dried on the glass microfiber and stored at room temperature for more than three months, after which it was still capable of detecting the target sequence. Finally, the feasibility of fiber-based FISH detection for genomic Legionella DNA was tested. Without further processing, the probe was unable to detect its target sequence in the complex genomic DNA. However, after selecting and application of appropriate restriction enzymes, specific detection of Legionella DNA against other aquatic pathogens with similar fragment patterns as Acinetobacter haemolyticus was possible.
Developmental Gains in Physical Fitness Components of Keyage and Older-than-Keyage Third-Graders
(2022)
Children who were enrolled according to legal enrollment dates (i.e., keyage third-graders aged eight to nine years) exhibit a positive linear physical fitness development (Fühner et al., 2021). However, children who were enrolled with a delay of one year or who repeated a grade (i.e., older-than-keyage children [OTK] aged nine to ten years in third grade) appear to exhibit a poorer physical fitness relative to what could be expected given their chronological age (Fühner et al., 2022). However, because Fühner et al. (2022) compared the performance of OTK children to predicted test scores that were extrapolated based on the data of keyage children, the observed physical fitness of these children could either indicate a delayed physical-fitness development or some physiological or psychological changes occurring during the tenth year of life. We investigate four hypotheses about this effect. (H1) OTK children are biologically younger than keyage children. A formula transforming OTK’s chronological age into a proxy for their biological age brings some of the observed cross-sectional age-related development in line with the predicted age-related development based on the data of keyage children, but large negative group differences remain. Hypotheses 2 to 4 were tested with a longitudinal assessment. (H2) Physiological changes due to biological maturation or psychological factors cause a stagnation of physical fitness development in the tenth year of life. H2 predicts a decline of performance from third to fourth grade also for keyage children. (H3) OTK children exhibit an age-related (temporary) developmental delay in the tenth year of life, but later catch up to the performance of age-matched keyage children. H3 predicts a larger developmental gain for OTK than for keyage children from third to fourth grade. (H4) OTK children exhibit a sustained physical fitness deficit and do not catch up over time. H4 predicts a positive development for keyage and OTK children, with no greater development for OTK compared to keyage children. The longitudinal study was based on a subset of children from the EMOTIKON project (www.uni-potsdam.de/emotikon). The physical fitness (cardiorespiratory endurance [6-minute-run test], coordination [star-run test], speed [20-m sprint test], lower [standing long jump test] and upper [ball push test] limbs muscle power, and balance [one-legged stance test]) of 1,274 children (1,030 keyage and 244 OTK children) from 32 different schools was tested in third grade and retested one year later in fourth grade. Results: (a) Both keyage and OTK children exhibit a positive longitudinal development from third to fourth grade in all six physical fitness components. (b) There is no evidence for a different longitudinal development of keyage and OTK children. (c) Keyage children (approximately 9.5 years in fourth grade) outperform age-matched OTK children (approximately 9.5 years in third grade) in all six physical fitness components. The results show that the physical fitness of OTK children is indeed impaired and are in support of a sustained difference in physical fitness between the groups of keyage and OTK children (H4).
Diet analysis of bats killed at wind turbines suggests large-scale losses of trophic interactions
(2022)
Agricultural practice has led to landscape simplification and biodiversity decline, yet recently, energy-producing infrastructures, such as wind turbines, have been added to these simplified agroecosystems, turning them into multi-functional energy-agroecosystems. Here, we studied the trophic interactions of bats killed at wind turbines using a DNA metabarcoding approach to shed light on how turbine-related bat fatalities may possibly affect local habitats. Specifically, we identified insect DNA in the stomachs of common noctule bats (Nyctalus noctula) killed by wind turbines in Germany to infer in which habitats these bats hunted. Common noctule bats consumed a wide variety of insects from different habitats, ranging from aquatic to terrestrial ecosystems (e.g., wetlands, farmland, forests, and grasslands). Agricultural and silvicultural pest insects made up about 20% of insect species consumed by the studied bats. Our study suggests that the potential damage of wind energy production goes beyond the loss of bats and the decline of bat populations. Bat fatalities at wind turbines may lead to the loss of trophic interactions and ecosystem services provided by bats, which may add to the functional simplification and impaired crop production, respectively, in multi-functional ecosystems.
Biomimicry is the art of mimicking nature to overcome a particular technical or scientific challenge. The approach studies how evolution has found solutions to the most complex problems in nature. This makes it a powerful method for science. In combination with the rapid development of manufacturing and information technologies into the digital age, structures and material that were before thought to be unrealizable can now be created with simple sketch and the touch of a button. This doctoral thesis had as its primary goal to investigate how digital tools, such as programming, modelling, 3D-Design tools and 3D-Printing, with the help from biomimicry, could lead to new analysis methods in science and new medical devices in medicine.
The Electrical Discharge Machining (EDM) process is applied commonly to deform or mold hard metals that are difficult to work using normal machinery. A workpiece submerged in an electrolyte is deformed while being in close vicinity to an electrode. When high voltage is put between the workpiece and the electrode it will cause sparks that create cavitations on the substrate which in turn removes material and is flushed away by the electrolyte. Usually, such surfaces are analysed based on roughness, in this work another method using a novel curvature analysis method is presented as an alternative. In addition, to better understand how the surface changes during process time of the EDM process, a digital impact model was created which created craters on ridges on an originally flat substrate. These substrates were then analysed using the curvature analysis method at different processing times of the modelling. It was found that a substrate reaches an equilibrium at around 10000 impacts. The proposed curvature analysis method has potential to be used in the design of new cell culture substrates for stem cell.
The Venus flytrap can shut its jaws at an amazing speed. The shutting mechanism may be interesting to use in science and is an example of a so-called mechanical bi-stable system – there are two stable states. In this work two truncated pyramid structures were modelled using a non-linear mechanical model called the Chained Beam Constraint Model (CBCM). The structure with a slope angle of 30 degrees is not bi-stable and the structure with a slope angle of 45 degrees is bi-stable. Developing this idea further by using PEVA, which has a shape-memory effect, the structure which is not bi-stable could be programmed to be bi-stable and then turned off again. This could be used as an energy storage system. Another species which has interesting mechanism is the tapeworm. Some species of this animal has a crown of hooks and suckers located on its side. The parasite commonly is found in mammals in the lower intestine and attaches to the walls by using its suckers. When the tapeworm has found a suitable spot, it ejects its hooks and permanently attaches to the wall. This function could be used in minimally invasive medicine to have better control of implants during the implantation process. By using the CBCM model and a 3D-printer capable of tuning how hard or soft a printed part is, a design strategy was developed to investigate how one could create a device that mimics the tapeworm. In the end a prototype was created which was able attach to a pork loin at an under pressure of 20 kPa and to ejects its hooks at an under pressure of 50 kPa or above.
These three projects is an exhibit of how digital tools and biomimicry can be used together to come up with applicable solutions in science and in medicine.
It is estimated that data scientists spend up to 80% of the time exploring, cleaning, and transforming their data. A major reason for that expenditure is the lack of knowledge about the used data, which are often from different sources and have heterogeneous structures. As a means to describe various properties of data, metadata can help data scientists understand and prepare their data, saving time for innovative and valuable data analytics. However, metadata do not always exist: some data file formats are not capable of storing them; metadata were deleted for privacy concerns; legacy data may have been produced by systems that were not designed to store and handle meta- data. As data are being produced at an unprecedentedly fast pace and stored in diverse formats, manually creating metadata is not only impractical but also error-prone, demanding automatic approaches for metadata detection.
In this thesis, we are focused on detecting metadata in CSV files – a type of plain-text file that, similar to spreadsheets, may contain different types of content at arbitrary positions. We propose a taxonomy of metadata in CSV files and specifically address the discovery of three different metadata: line and cell type, aggregations, and primary keys and foreign keys.
Data are organized in an ad-hoc manner in CSV files, and do not follow a fixed structure, which is assumed by common data processing tools. Detecting the structure of such files is a prerequisite of extracting information from them, which can be addressed by detecting the semantic type, such as header, data, derived, or footnote, of each line or each cell. We propose the supervised- learning approach Strudel to detect the type of lines and cells. CSV files may also include aggregations. An aggregation represents the arithmetic relationship between a numeric cell and a set of other numeric cells. Our proposed AggreCol algorithm is capable of detecting aggregations of five arithmetic functions in CSV files. Note that stylistic features, such as font style and cell background color, do not exist in CSV files. Our proposed algorithms address the respective problems by using only content, contextual, and computational features.
Storing a relational table is also a common usage of CSV files. Primary keys and foreign keys are important metadata for relational databases, which are usually not present for database instances dumped as plain-text files. We propose the HoPF algorithm to holistically detect both constraints in relational databases. Our approach is capable of distinguishing true primary and foreign keys from a great amount of spurious unique column combinations and inclusion dependencies, which can be detected by state-of-the-art data profiling algorithms.
Botulinum neurotoxin (BoNT) is used for the treatment of a number of ailments. The activity of the toxin that is isolated from bacterial cultures is frequently tested in the mouse lethality assay. Apart from the ethical concerns inherent to this assay, species-specific differences in the affinity for different BoNT serotypes give rise to activity results that differ from the activity in humans. Thus, BoNT/B is more active in mice than in humans. The current study shows that the stimulus-dependent release of a luciferase from a differentiated human neuroblastoma–based reporter cell line (SIMA-hPOMC1-26-Gluc) was inhibited by clostridial and recombinant BoNT/A to the same extent, whereas both clostridial and recombinant BoNT/B inhibited the release to a lesser extent and only at much higher concentrations, reflecting the low activity of BoNT/B in humans. By contrast, the genetically modified BoNT/B-MY, which has increased affinity for human synaptotagmin, and the BoNT/B protein receptor inhibited luciferase release effectively and with an EC50 comparable to recombinant BoNT/A. This was due to an enhanced uptake into the reporter cells of BoNT/B-MY in comparison to the recombinant wild-type toxin. Thus, the SIMA-hPOMC1-26-Gluc cell assay is a versatile tool to determine the activity of different BoNT serotypes providing human-relevant dose-response data.
Infectious diseases are an increasing threat to biodiversity and human health. Therefore, developing a general understanding of the drivers shaping host-pathogen dynamics is of key importance in both ecological and epidemiological research. Disease dynamics are driven by a variety of interacting processes such as individual host behaviour, spatiotemporal resource availability or pathogen traits like virulence and transmission. External drivers such as global change may modify the system conditions and, thus, the disease dynamics. Despite their importance, many of these drivers are often simplified and aggregated in epidemiological models and the interactions among multiple drivers are neglected.
In my thesis, I investigate disease dynamics using a mechanistic approach that includes both bottom-up effects - from landscape dynamics to individual movement behaviour - as well as top-down effects - from pathogen virulence on host density and contact rates. To this end, I extended an established spatially explicit individual-based model that simulates epidemiological and ecological processes stochastically, to incorporate a dynamic resource landscape that can be shifted away from the timing of host population-dynamics (chapter 2). I also added the evolution of pathogen virulence along a theoretical virulence-transmission trade-off (chapter 3). In chapter 2, I focus on bottom-up effects, specifically how a temporal shift of resource availability away from the timing of biological events of host-species - as expected under global change - scales up to host-pathogen interactions and disease dynamics. My results show that the formation of temporary disease hotspots in combination with directed individual movement acted as key drivers for pathogen persistence even under highly unfavourable conditions for the host. Even with drivers like global change further increasing the likelihood of unfavourable interactions between host species and their environment, pathogens can continue to persist with heir hosts. In chapter 3, I demonstrate that the top-down effect caused by pathogen-associated mortality on its host population can be mitigated by selection for lower virulent pathogen strains when host densities are reduced through mismatches between seasonal resource availability and host life-history events. I chapter 4, I combined parts of both theoretical models into a new model that includes individual host movement decisions and the evolution of pathogenic virulence to simulate pathogen outbreaks in realistic landscapes. I was able to match simulated patterns of pathogen spread to observed patterns from long-term outbreak data of classical swine fever in wild boar in Northern Germany. The observed disease course was best explained by a simulated high virulent strain, whereas sampling schemes and vaccination campaigns could explain differences in the age-distribution of infected hosts. My model helps to understand and disentangle how the combination of individual decision making and evolution of virulence can act as important drivers of pathogen spread and persistence.
As I show across the chapters of this thesis, the interplay of both bottom-up and top-down processes is a key driver of disease dynamics in spatially structured host populations, as they ultimately shape host densities and contact rates among moving individuals. My findings are an important step towards a paradigm shift in disease ecology away from simplified assumptions towards the inclusion of mechanisms, such as complex multi-trophic interactions, and their feedbacks on pathogen spread and disease persistence. The mechanisms presented here should be at the core of realistic predictive and preventive epidemiological models.
In this paper we examine the effect of uncertainty on readers’ predictions about meaning. In particular, we were interested in how uncertainty might influence the likelihood of committing to a specific sentence meaning. We conducted two event-related potential (ERP) experiments using particle verbs such as turn down and manipulated uncertainty by constraining the context such that readers could be either highly certain about the identity of a distant verb particle, such as turn the bed […] down, or less certain due to competing particles, such as turn the music […] up/down. The study was conducted in German, where verb particles appear clause-finally and may be separated from the verb by a large amount of material. We hypothesised that this separation would encourage readers to predict the particle, and that high certainty would make prediction of a specific particle more likely than lower certainty. If a specific particle was predicted, this would reflect a strong commitment to sentence meaning that should incur a higher processing cost if the prediction is wrong. If a specific particle was less likely to be predicted, commitment should be weaker and the processing cost of a wrong prediction lower. If true, this could suggest that uncertainty discourages predictions via an unacceptable cost-benefit ratio. However, given the clear predictions made by the literature, it was surprisingly unclear whether the uncertainty manipulation affected the two ERP components studied, the N400 and the PNP. Bayes factor analyses showed that evidence for our a priori hypothesised effect sizes was inconclusive, although there was decisive evidence against a priori hypothesised effect sizes larger than 1μV for the N400 and larger than 3μV for the PNP. We attribute the inconclusive finding to the properties of verb-particle dependencies that differ from the verb-noun dependencies in which the N400 and PNP are often studied.
Background: Children’s spontaneous focusing on numerosity (SFON) is related to numerical skills. This study aimed to examine (1) the developmental trajectory of SFON and (2) the interrelations between SFON and early numerical skills at pre-school as well as their influence on arithmetical skills at school. Method: Overall, 1868 German pre-school children were repeatedly assessed until second grade. Nonverbal intelligence, visual attention, visuospatial working memory, SFON and numerical skills were assessed at age five (M = 63 months, Time 1) and age six (M = 72 months, Time 2), and arithmetic was assessed at second grade (M = 95 months, Time 3). Results: SFON increased significantly during pre-school. Path analyses revealed interrelations between SFON and several numerical skills, except number knowledge. Magnitude estimation and basic calculation skills (Time 1 and Time 2), and to a small degree number knowledge (Time 2), contributed directly to arithmetic in second grade. The connection between SFON and arithmetic was fully mediated by magnitude estimation and calculation skills at pre-school. Conclusion: Our results indicate that SFON first and foremost influences deeper understanding of numerical concepts at pre-school and—in contrast to previous findings –affects only indirectly children’s arithmetical development at school.
Doing good by doing bad
(2022)
This study investigates how tone at the top, implemented by top management, and tone at the bottom, in an employee's immediate work environment, determine noncompliance. We focus on the disallowed actions of employees that improve their own and, in turn, the company's performance, referred to as performance-improving noncompliant behavior (PINC behavior). We conduct a survey of German sales employees to investigate specifically how, on the one hand, (1) corporate rules and (2) performance pressure, both implemented by top management, and, on the other hand, (3) others' PINC expectations and (4) others' PINC behavior, both arising from the employee's immediate work environment, influence PINC behavior. When considered in isolation, we find that corporate rules, as top management's main instrument to guide employee behavior, decrease employee PINC behavior. However, this effect is negatively influenced by the employees' immediate work environment when employees are expected to engage in PINC or when others engage in PINC. In contrast, even though top management places great performance pressure on employees, that by itself does not increase PINC behavior. Overall, our study informs practitioners and researchers about whether and how the four determinants increase or decrease employees' PINC behavior, which is important to comprehend triggers and to counteract such misconduct.
In semi-arid environments characterized by erratic rainfall and scattered primary production, migratory movements are a key survival strategy of large herbivores to track resources over vast areas. Veterinary Cordon Fences (VCFs), intended to reduce wildlife-livestock disease transmission, fragment large parts of southern Africa and have limited the movements of large wild mammals for over 60 years. Consequently, wildlife-fence interactions are frequent and often result in perforations of the fence, mainly caused by elephants. Yet, we lack knowledge about at which times fences act as barriers, how fences directly alter the energy expenditure of native herbivores, and what the consequences of impermeability are. We studied 2-year ungulate movements in three common antelopes (springbok, kudu, eland) across a perforated part of Namibia's VCF separating a wildlife reserve and Etosha National Park using GPS telemetry, accelerometer measurements, and satellite imagery. We identified 2905 fence interaction events which we used to evaluate critical times of encounters and direct fence effects on energy expenditure. Using vegetation type-specific greenness dynamics, we quantified what animals gained in terms of high quality food resources from crossing the VCF. Our results show that the perforation of the VCF sustains herbivore-vegetation interactions in the savanna with its scattered resources. Fence permeability led to peaks in crossing numbers during the first flush of woody plants before the rain started. Kudu and eland often showed increased energy expenditure when crossing the fence. Energy expenditure was lowered during the frequent interactions of ungulates standing at the fence. We found no alteration of energy expenditure when springbok immediately found and crossed fence breaches. Our results indicate that constantly open gaps did not affect energy expenditure, while gaps with obstacles increased motion. Closing gaps may have confused ungulates and modified their intended movements. While browsing, sedentary kudu's use of space was less affected by the VCF; migratory, mixed-feeding springbok, and eland benefited from gaps by gaining forage quality and quantity after crossing. This highlights the importance of access to vast areas to allow ungulates to track vital vegetation patches.
An Arabidopsis mutant lacking both the cytosolic Disproportionating enzyme 2 (DPE2) and the plastidial glucan Phosphorylase 1 (PHS1) revealed a unique starch metabolism. Dpe2/phs1 has been reported to have only one starch granule number per chloroplast when grown under diurnal rhythm. For this study, we analyzed dpe2/phs1 in details following the mutant development, and found that it showed three distinct periods of granule numbers per chloroplast, while there was no obvious change observed in Col-0. In young plants, the starch granule number was similar to that in Col-0 at first, and then decreased significantly, down to one or no granule per chloroplast, followed by an increase in the granule number. Thus, in dpe2/phs1, control over the starch granule number is impaired, but it is not defective in starch granule initiation. The data also indicate that the granule number is not fixed, and is regulated throughout plant growth. Furthermore, the chloroplasts revealed alterations during these three periods, with a partially strong aberrant morphology in the middle phase. Interestingly, the unique metabolism was perpetuated when starch degradation was further impaired through an additional lack of Isoamylase 3 (ISA3) or Starch excess 4 (SEX4). Transcriptomic studies and metabolic profiling revealed the co-regulation of starch metabolism-related genes and a clear metabolic separation between the periods. Most senescence-induced genes were found to be up-regulated more than twice in the starch-less mature leaves. Thus, dpe2/phs1 is a unique plant material source, with which we may study starch granule number regulation to obtain a more detailed understanding.
Isoflux tension propagation (IFTP) theory and Langevin dynamics (LD) simulations are employed to study the dynamics of channel-driven polymer translocation in which a polymer translocates into a narrow channel and the monomers in the channel experience a driving force fc. In the high driving force limit, regardless of the channel width, IFTP theory predicts τ ∝ f βc for the translocation time, where β = −1 is the force scaling exponent. Moreover, LD data show that for a very narrow channel fitting only a single file of monomers, the entropic force due to the subchain inside the channel does not play a significant role in the translocation dynamics and the force exponent β = −1 regardless of the force magnitude. As the channel width increases the number of possible spatial configurations of the subchain inside the channel becomes significant and the resulting entropic force causes the force exponent to drop below unity.
The Arctic is greatly affected by climate change. Increasing air temperatures drive permafrost thaw and an increase in coastal erosion and river discharge. This results in a greater input of sediment and organic matter into nearshore waters, impacting ecosystems by reducing light transmission through the water column and altering biogeochemistry. This potentially results in impacts on the subsistence economy of local people as well as the climate due to the transformation of suspended organic matter into greenhouse gases. Even though the impacts of increased suspended sediment concentrations and turbidity in the Arctic nearshore zone are well-studied, the mechanisms underpinning this increase are largely unknown. Wave energy and tides drive the level of turbidity in the temperate and tropical parts of the world, and this is generally assumed to also be the case in the Arctic. However, the tidal range is considerably lower in the Arctic, and processes related to the occurrence of permafrost have the potential to greatly contribute to nearshore turbidity. In this study, we use high-resolution satellite imagery alongside in situ and ERA5 reanalysis data of ocean and climate variables in order to identify the drivers of nearshore turbidity, along with its seasonality in the nearshore waters of Herschel Island Qikiqtaruk, in the western Canadian Arctic. Nearshore turbidity correlates well to wind direction, wind speed, significant wave height, and wave period. Nearshore turbidity is superiorly correlated to wind speed at the Beaufort Shelf compared to in situ measurements at Herschel Island Qikiqtaruk, showing that nearshore turbidity, albeit being of limited spatial extent, is influenced by large-scale weather and ocean phenomenons. We show that, in contrast to the temperate and tropical ocean, freshly eroded material is the predominant driver of nearshore turbidity in the Arctic, rather than resuspension, which is caused by the vulnerability of permafrost coasts to thermo-erosion.
Peripersonal space is the space surrounding our body, where multisensory integration of stimuli and action execution take place. The size of peripersonal space is flexible and subject to change by various personal and situational factors. The dynamic representation of our peripersonal space modulates our spatial behaviors towards other individuals. During the COVID-19 pandemic, this spatial behavior was modified by two further factors: social distancing and wearing a face mask. Evidence from offline and online studies on the impact of a face mask on pro-social behavior is mixed. In an attempt to clarify the role of face masks as pro-social or anti-social signals, 235 observers participated in the present online study. They watched pictures of two models standing at three different distances from each other (50, 90 and 150 cm), who were either wearing a face mask or not and were either interacting by initiating a hand shake or just standing still. The observers’ task was to classify the model by gender. Our results show that observers react fastest, and therefore show least avoidance, for the shortest distances (50 and 90 cm) but only when models wear a face mask and do not interact. Thus, our results document both pro- and anti-social consequences of face masks as a result of the complex interplay between social distancing and interactive behavior. Practical implications of these findings are discussed.
The deciduous needle tree larch (Larix Mill.) covers more than 80% of the Asian boreal forests. Only a few Larix species constitute the vast forests and these species differ markedly in their ecological traits, most importantly in their ability to grow on and stabilize underlying permafrost. The pronounced dominance of the summergreen larches makes the Asian boreal forests unique, as the rest of the northern hemisphere boreal forests is almost exclusively dominated by evergreen needle-leaf forests. Global warming is impacting the whole world but is especially pronounced in the arctic and boreal regions. Although adapted to extreme climatic conditions, larch forests are sensitive to varying climatic conditions. By their sheer size, changes in Asian larch forests as range shifts or changes in species composition and the resulting vegetation-climate feedbacks are of global relevance. It is however still uncertain if larch forests will persist under the ongoing warming climate or if they will be replaced by evergreen forests. It is therefore of great importance to understand how these ecosystems will react to future climate warmings and if they will maintain their dominance. One step in the better understanding of larch dynamics is to study how the vast dominant forests developed and why they only established in northern Asia. A second step is to study how the species reacted to past changes in the climate.
The first objective of this thesis was to review and identify factors promoting Asian larch dominance. I achieved this by synthesizing and comparing reported larch occurrences and influencing components on the northern hemisphere continents in the present and in the past. The second objective was to find a possibility to directly study past Larix populations in Siberia and specifically their genetic variation, enabling the study of geographic movements. For this, I established chloroplast enrichment by hybridization capture from sedimentary ancient DNA (sedaDNA) isolated from lake sediment records. The third objective was to use the established method to track past larch populations, their glacial refugia during the Last Glacial Maximum (LGM) around 21,000 years before present (ka BP), and their post-glacial migration patterns.
To study larch promoting factors, I compared the present state of larch species ranges, areas of dominance, their bioclimatic niches, and the distribution on different extents and thaw depths of permafrost. The species comparison showed that the bioclimatic niches greatly overlap between the American and Asian species and that it is only in the extremely continental climates in which only the Asian larch species can persist. I revealed that the area of dominance is strongly connected to permafrost extent but less linked to permafrost seasonal thaw depths. Comparisons of the paleorecord of larch between the continents suggest differences in the recolonization history. Outside of northern Asia and Alaska, glacial refugial populations of larch were confined to the southern regions and thus recolonization could only occur as migration from south to north. Alaskan larch populations could not establish wide-range dominant forest which could be related to their own genetically depletion as separated refugial population. In Asia, it is still unclear whether or not the northern refugial populations contributed and enhanced the postglacial colonization or whether they were replaced by populations invading from the south in the course of climate warming. Asian larch dominance is thus promoted partly by adaptions to extremely continental climates and by adaptations to grow on continuous permafrost but could be also connected to differences in glacial survival and recolonization history of Larix species.
Except for extremely rare macrofossil findings of fossilized cones, traditional methods to study past vegetation are not able to distinguish between larch species or populations. Within the scope of this thesis, I therefore established a method to retrieve genetic information of past larch populations to distinguish between species. Using the Larix chloroplast genome as target, I successfully applied the method of DNA target enrichment by hybridization capture on sedaDNA samples from lake records and showed that it is able to distinguish between larch species. I then used the method on samples from lake records from across Siberia dating back up to 50 ka BP. The results allowed me to address the question of glacial survival and post-glacial recolonization mode in Siberian larch species. The analyzed pattern showed that LGM refugia were almost exclusively constituted by L. gmelinii, even in sites of current L. sibirica distribution. For included study sites, L. sibirica migrated into its extant northern distribution area only in the Holocene. Consequently, the post-glacial recolonization of L. sibirica was not enhanced by northern glacial refugia. In case of sites in extant distribution area of L. gmelinii, the absence of a genetic turn-over point to a continuous population rather than an invasion of southern refugia. The results suggest that climate has a strong influence on the distribution of Larix species and that species may also respond differently to future climate warming. Because species differ in their ecological characteristics, species distribution is also relevant with respect to further feedbacks between vegetation and climate.
With this thesis, I give an overview of present and past larch occurrences and evaluate which factors promote their dominance. Furthermore, I provide the tools to study past Larix species and give first important insights into the glacial history of Larix populations.
Objective: To examine the effect of plyometric jump training on skeletal muscle hypertrophy in healthy individuals.
Methods: A systematic literature search was conducted in the databases PubMed, SPORTDiscus, Web of Science, and Cochrane Library up to September 2021.
Results: Fifteen studies met the inclusion criteria. The main overall finding (44 effect sizes across 15 clusters median = 2, range = 1–15 effects per cluster) indicated that plyometric jump training had small to moderate effects [standardised mean difference (SMD) = 0.47 (95% CIs = 0.23–0.71); p < 0.001] on skeletal muscle hypertrophy. Subgroup analyses for training experience revealed trivial to large effects in non-athletes [SMD = 0.55 (95% CIs = 0.18–0.93); p = 0.007] and trivial to moderate effects in athletes [SMD = 0.33 (95% CIs = 0.16–0.51); p = 0.001]. Regarding muscle groups, results showed moderate effects for the knee extensors [SMD = 0.72 (95% CIs = 0.66–0.78), p < 0.001] and equivocal effects for the plantar flexors [SMD = 0.65 (95% CIs = −0.25–1.55); p = 0.143]. As to the assessment methods of skeletal muscle hypertrophy, findings indicated trivial to small effects for prediction equations [SMD = 0.29 (95% CIs = 0.16–0.42); p < 0.001] and moderate-to-large effects for ultrasound imaging [SMD = 0.74 (95% CIs = 0.59–0.89); p < 0.001]. Meta-regression analysis indicated that the weekly session frequency moderates the effect of plyometric jump training on skeletal muscle hypertrophy, with a higher weekly session frequency inducing larger hypertrophic gains [β = 0.3233 (95% CIs = 0.2041–0.4425); p < 0.001]. We found no clear evidence that age, sex, total training period, single session duration, or the number of jumps per week moderate the effect of plyometric jump training on skeletal muscle hypertrophy [β = −0.0133 to 0.0433 (95% CIs = −0.0387 to 0.1215); p = 0.101–0.751].
Conclusion: Plyometric jump training can induce skeletal muscle hypertrophy, regardless of age and sex. There is evidence for relatively larger effects in non-athletes compared with athletes. Further, the weekly session frequency seems to moderate the effect of plyometric jump training on skeletal muscle hypertrophy, whereby more frequent weekly plyometric jump training sessions elicit larger hypertrophic adaptations.
The protein fraction, important for coffee cup quality, is modified during post-harvest treatment prior to roasting. Proteins may interact with phenolic compounds, which constitute the major metabolites of coffee, where the processing affects these interactions. This allows the hypothesis that the proteins are denatured and modified via enzymatic and/or redox activation steps. The present study was initiated to encompass changes in the protein fraction. The investigations were limited to major storage protein of green coffee beans. Fourteen Coffea arabica samples from various processing methods and countries were used. Different extraction protocols were compared to maintain the status quo of the protein modification. The extracts contained about 4–8 µg of chlorogenic acid derivatives per mg of extracted protein. High-resolution chromatography with multiple reaction monitoring was used to detect lysine modifications in the coffee protein. Marker peptides were allocated for the storage protein of the coffee beans. Among these, the modified peptides K.FFLANGPQQGGK.E and R.LGGK.T of the α-chain and R.ITTVNSQK.I and K.VFDDEVK.Q of β-chain were detected. Results showed a significant increase (p < 0.05) of modified peptides from wet processed green beans as compared to the dry ones. The present study contributes to a better understanding of the influence of the different processing methods on protein quality and its role in the scope of coffee cup quality and aroma. View Full-Text
Timing of initial school enrollment may vary considerably for various reasons such as early or delayed enrollment, skipped or repeated school classes. Accordingly, the age range within school grades includes older-(OTK) and younger-than-keyage (YTK) children. Hardly any information is available on the impact of timing of school enrollment on physical fitness. There is evidence from a related research topic showing large differences in academic performance between OTK and YTK children versus keyage children. Thus, the aim of this study was to compare physical fitness of OTK (N = 26,540) and YTK (N = 2586) children versus keyage children (N = 108,295) in a representative sample of German third graders. Physical fitness tests comprised cardiorespiratory endurance, coordination, speed, lower, and upper limbs muscle power. Predictions of physical fitness performance for YTK and OTK children were estimated using data from keyage children by taking age, sex, school, and assessment year into account. Data were annually recorded between 2011 and 2019. The difference between observed and predicted z-scores yielded a delta z-score that was used as a dependent variable in the linear mixed models. Findings indicate that OTK children showed poorer performance compared to keyage children, especially in coordination, and that YTK children outperformed keyage children, especially in coordination. Teachers should be aware that OTK children show poorer physical fitness performance compared to keyage children.
Effects of fences and fence gaps on the movement behavior of three southern African antelope species
(2022)
Globally, migratory ungulates are affected by fences. While field observational studies reveal the amount of animal–fence interactions across taxa, GPS tracking-based studies uncover fence effects on movement patterns and habitat selection. However, studies on the direct effects of fences and fence gaps on movement behavior, especially based on high-frequency tracking data, are scarce. We used GPS tracking on three common African antelopes (Tragelaphus strepsiceros, Antidorcas marsupialis, and T. oryx) with movement strategies ranging from range residency to nomadism in a semi-arid, Namibian savanna traversed by wildlife-proof fences that elephants have regularly breached. We classified major forms of ungulate–fence interaction types on a seasonal and a daily scale. Furthermore, we recorded the distances and times spent at fences regarding the total individual space use. Based on this, we analyzed the direct effects of fences and fence gaps on the animals’ movement behavior for the previously defined types of animal–fence interactions. Antelope-fence interactions peaked during the early hours of the day and during seasonal transitions when the limiting resource changed between water and forage. Major types of ungulate–fence interactions were quick, trace-like, or marked by halts. We found that the amount of time spent at fences was highest for nomadic eland. Migratory springbok adjusted their space use concerning fence gap positions. If the small home ranges of sedentary kudu included a fence, they frequently interacted with this fence. For springbok and eland, distance traveled along a fence declined with increasing utilization of a fence gap. All species reduced their speed in the proximity of a fence but often increased their speed when encountering the fence. Crossing a fence led to increased speeds for all species. We demonstrate that fence effects mainly occur during crucial foraging times (seasonal scale) and during times of directed movements (daily scale). Importantly, we provide evidence that fences directly alter antelope movement behaviors with negative implications for energy budgets and that persistent fence gaps can reduce the intensity of such alterations. Our findings help to guide future animal–fence studies and provide insights for wildlife fencing and fence gap planning.
Physical fatigue (PF) negatively affects postural control, resulting in impaired balance performance in young and older adults. Similar effects on postural control can be observed for mental fatigue (MF) mainly in older adults. Controversial results exist for young adults. There is a void in the literature on the effects of fatigue on balance and cortical activity. Therefore, this study aimed to examine the acute effects of PF and MF on postural sway and cortical activity. Fifteen healthy young adults aged 28 ± 3 years participated in this study. MF and PF protocols comprising of an all-out repeated sit-to-stand task and a computer-based attention network test, respectively, were applied in random order. Pre and post fatigue, cortical activity and postural sway (i.e., center of pressure displacements [CoPd], velocity [CoPv], and CoP variability [CV CoPd, CV CoPv]) were tested during a challenging bipedal balance board task. Absolute spectral power was calculated for theta (4–7.5 Hz), alpha-2 (10.5–12.5 Hz), beta-1 (13–18 Hz), and beta-2 (18.5–25 Hz) in frontal, central, and parietal regions of interest (ROI) and baseline-normalized. Inference statistics revealed a significant time-by-fatigue interaction for CoPd (p = 0.009, d = 0.39, Δ 9.2%) and CoPv (p = 0.009, d = 0.36, Δ 9.2%), and a significant main effect of time for CoP variability (CV CoPd: p = 0.001, d = 0.84; CV CoPv: p = 0.05, d = 0.62). Post hoc analyses showed a significant increase in CoPd (p = 0.002, d = 1.03) and CoPv (p = 0.003, d = 1.03) following PF but not MF. For cortical activity, a significant time-by-fatigue interaction was found for relative alpha-2 power in parietal (p < 0.001, d = 0.06) areas. Post hoc tests indicated larger alpha-2 power increases after PF (p < 0.001, d = 1.69, Δ 3.9%) compared to MF (p = 0.001, d = 1.03, Δ 2.5%). In addition, changes in parietal alpha-2 power and measures of postural sway did not correlate significantly, irrespective of the applied fatigue protocol. No significant changes were found for the other frequency bands, irrespective of the fatigue protocol and ROI under investigation. Thus, the applied PF protocol resulted in increased postural sway (CoPd and CoPv) and CoP variability accompanied by enhanced alpha-2 power in the parietal ROI while MF led to increased CoP variability and alpha-2 power in our sample of young adults. Potential underlying cortical mechanisms responsible for the greater increase in parietal alpha-2 power after PF were discussed but could not be clearly identified as cause. Therefore, further future research is needed to decipher alternative interpretations.
Background
A growing body of literature is available regarding the effects of plyometric jump training (PJT) on measures of physical fitness (PF) and sport-specific performance (SSP) in-water sports athletes (WSA, i.e. those competing in sports that are practiced on [e.g. rowing] or in [e.g. swimming; water polo] water). Indeed, incoherent findings have been observed across individual studies making it difficult to provide the scientific community and coaches with consistent evidence. As such, a comprehensive systematic literature search should be conducted to clarify the existent evidence, identify the major gaps in the literature, and offer recommendations for future studies.
Aim
To examine the effects of PJT compared with active/specific-active controls on the PF (one-repetition maximum back squat strength, squat jump height, countermovement jump height, horizontal jump distance, body mass, fat mass, thigh girth) and SSP (in-water vertical jump, in-water agility, time trial) outcomes in WSA, through a systematic review with meta-analysis of randomized and non-randomized controlled studies.
Methods
The electronic databases PubMed, Scopus, and Web of Science were searched up to January 2022. According to the PICOS approach, the eligibility criteria were: (population) healthy WSA; (intervention) PJT interventions involving unilateral and/or bilateral jumps, and a minimal duration of ≥ 3 weeks; (comparator) active (i.e. standard sports training) or specific-active (i.e. alternative training intervention) control group(s); (outcome) at least one measure of PF (e.g. jump height) and/or SSP (e.g. time trial) before and after training; and (study design) multi-groups randomized and non-randomized controlled trials. The Physiotherapy Evidence Database (PEDro) scale was used to assess the methodological quality of the included studies. The DerSimonian and Laird random-effects model was used to compute the meta-analyses, reporting effect sizes (ES, i.e. Hedges’ g) with 95% confidence intervals (95% CIs). Statistical significance was set at p ≤ 0.05. Certainty or confidence in the body of evidence for each outcome was assessed using Grading of Recommendations Assessment, Development, and Evaluation (GRADE), considering its five dimensions: risk of bias in studies, indirectness, inconsistency, imprecision, and risk of publication bias.
Results
A total of 11,028 studies were identified with 26 considered eligible for inclusion. The median PEDro score across the included studies was 5.5 (moderate-to-high methodological quality). The included studies involved a total of 618 WSA of both sexes (330 participants in the intervention groups [31 groups] and 288 participants in the control groups [26 groups]), aged between 10 and 26 years, and from different sports disciplines such as swimming, triathlon, rowing, artistic swimming, and water polo. The duration of the training programmes in the intervention and control groups ranged from 4 to 36 weeks. The results of the meta-analysis indicated no effects of PJT compared to control conditions (including specific-active controls) for in-water vertical jump or agility (ES = − 0.15 to 0.03; p = 0.477 to 0.899), or for body mass, fat mass, and thigh girth (ES = 0.06 to 0.15; p = 0.452 to 0.841). In terms of measures of PF, moderate-to-large effects were noted in favour of the PJT groups compared to the control groups (including specific-active control groups) for one-repetition maximum back squat strength, horizontal jump distance, squat jump height, and countermovement jump height (ES = 0.67 to 1.47; p = 0.041 to < 0.001), in addition to a small effect noted in favour of the PJT for SSP time-trial speed (ES = 0.42; p = 0.005). Certainty of evidence across the included studies varied from very low-to-moderate.
Conclusions
PJT is more effective to improve measures of PF and SSP in WSA compared to control conditions involving traditional sport-specific training as well as alternative training interventions (e.g. resistance training). It is worth noting that the present findings are derived from 26 studies of moderate-to-high methodological quality, low-to-moderate impact of heterogeneity, and very low-to-moderate certainty of evidence based on GRADE.
Trial registration The protocol for this systematic review with meta-analysis was published in the Open Science platform (OSF) on January 23, 2022, under the registration doi https://doi.org/10.17605/OSF.IO/NWHS3 (internet archive link: https://archive.org/details/osf-registrations-nwhs3-v1).
Background: In terms of physiological and biomechanical characteristics, over-pronation of the feet has been associated with distinct muscle recruitment patterns and ground reaction forces during running.
Objective: The aim of this study was to evaluate the effects of running on sand vs. stable ground on ground-reaction-forces (GRFs) and electromyographic (EMG) activity of lower limb muscles in individuals with over-pronated feet (OPF) compared with healthy controls.
Methods: Thirty-three OPF individuals and 33 controls ran at preferred speed and in randomized-order over level-ground and sand. A force-plate was embedded in an 18-m runway to collect GRFs. Muscle activities were recorded using an EMG-system. Data were adjusted for surface-related differences in running speed.
Results: Running on sand resulted in lower speed compared with stable ground running (p < 0.001; d = 0.83). Results demonstrated that running on sand produced higher tibialis anterior activity (p = 0.024; d = 0.28). Also, findings indicated larger loading rates (p = 0.004; d = 0.72) and greater vastus medialis (p < 0.001; d = 0.89) and rectus femoris (p = 0.001; d = 0.61) activities in OPF individuals. Controls but not OPF showed significantly lower gluteus-medius activity (p = 0.022; d = 0.63) when running on sand.
Conclusion: Running on sand resulted in lower running speed and higher tibialis anterior activity during the loading phase. This may indicate alterations in neuromuscular demands in the distal part of the lower limbs when running on sand. In OPF individuals, higher loading rates together with greater quadriceps activity may constitute a proximal compensatory mechanism for distal surface instability.
Aims: High intensity interval training (HIIT) improves mitochondrial characteristics. This study compared the impact of two workload-matched high intensity interval training (HIIT) protocols with different work:recovery ratios on regulatory factors related to mitochondrial biogenesis in the soleus muscle of diabetic rats.
Materials and methods: Twenty-four Wistar rats were randomly divided into four equal-sized groups: non-diabetic control, diabetic control (DC), diabetic with long recovery exercise [4–5 × 2-min running at 80%–90% of the maximum speed reached with 2-min of recovery at 40% of the maximum speed reached (DHIIT1:1)], and diabetic with short recovery exercise (5–6 × 2-min running at 80%–90% of the maximum speed reached with 1-min of recovery at 30% of the maximum speed reached [DHIIT2:1]). Both HIIT protocols were completed five times/week for 4 weeks while maintaining equal running distances in each session.
Results: Gene and protein expressions of PGC-1α, p53, and citrate synthase of the muscles increased significantly following DHIIT1:1 and DHIIT2:1 compared to DC (p ˂ 0.05). Most parameters, except for PGC-1α protein (p = 0.597), were significantly higher in DHIIT2:1 than in DHIIT1:1 (p ˂ 0.05). Both DHIIT groups showed significant increases in maximum speed with larger increases in DHIIT2:1 compared with DHIIT1:1.
Conclusion: Our findings indicate that both HIIT protocols can potently up-regulate gene and protein expression of PGC-1α, p53, and CS. However, DHIIT2:1 has superior effects compared with DHIIT1:1 in improving mitochondrial adaptive responses in diabetic rats.
Duplicate detection describes the process of finding multiple representations of the same real-world entity in the absence of a unique identifier, and has many application areas, such as customer relationship management, genealogy and social sciences, or online shopping. Due to the increasing amount of data in recent years, the problem has become even more challenging on the one hand, but has led to a renaissance in duplicate detection research on the other hand.
This thesis examines the effects and opportunities of transitive relationships on the duplicate detection process. Transitivity implies that if record pairs ⟨ri,rj⟩ and ⟨rj,rk⟩ are classified as duplicates, then also record pair ⟨ri,rk⟩ has to be a duplicate. However, this reasoning might contradict with the pairwise classification, which is usually based on the similarity of objects. An essential property of similarity, in contrast to equivalence, is that similarity is not necessarily transitive.
First, we experimentally evaluate the effect of an increasing data volume on the threshold selection to classify whether a record pair is a duplicate or non-duplicate. Our experiments show that independently of the pair selection algorithm and the used similarity measure, selecting a suitable threshold becomes more difficult with an increasing number of records due to an increased probability of adding a false duplicate to an existing cluster. Thus, the best threshold changes with the dataset size, and a good threshold for a small (possibly sampled) dataset is not necessarily a good threshold for a larger (possibly complete) dataset. As data grows over time, earlier selected thresholds are no longer a suitable choice, and the problem becomes worse for datasets with larger clusters.
Second, we present with the Duplicate Count Strategy (DCS) and its enhancement DCS++ two alternatives to the standard Sorted Neighborhood Method (SNM) for the selection of candidate record pairs. DCS adapts SNMs window size based on the number of detected duplicates and DCS++ uses transitive dependencies to save complex comparisons for finding duplicates in larger clusters. We prove that with a proper (domain- and data-independent!) threshold, DCS++ is more efficient than SNM without loss of effectiveness.
Third, we tackle the problem of contradicting pairwise classifications. Usually, the transitive closure is used for pairwise classifications to obtain a transitively closed result set. However, the transitive closure disregards negative classifications. We present three new and several existing clustering algorithms and experimentally evaluate them on various datasets and under various algorithm configurations. The results show that the commonly used transitive closure is inferior to most other clustering algorithms, especially for the precision of results. In scenarios with larger clusters, our proposed EMCC algorithm is, together with Markov Clustering, the best performing clustering approach for duplicate detection, although its runtime is longer than Markov Clustering due to the subexponential time complexity. EMCC especially outperforms Markov Clustering regarding the precision of the results and additionally has the advantage that it can also be used in scenarios where edge weights are not available.
Objective: The behaviors of endothelial cells or mesenchymal stem cells are remarkably influenced by the mechanical properties of their surrounding microenvironments. Here, electrospun fiber meshes containing various mechanical characteristics were developed from polyetheresterurethane (PEEU) copolymers. The goal of this study was to explore how fiber mesh stiffness affected endothelial cell shape, growth, migration, and angiogenic potential of endothelial cells. Furthermore, the effects of the E-modulus of fiber meshes on human adipose-derived stem cells (hADSCs) osteogenic potential was investigated.
Methods: Polyesteretherurethane (PEEU) polymers with various poly(p-dioxanone) (PPDO) to poly (ε-caprolactone) (PCL) weight percentages (40 wt.%, 50 wt.%, 60 wt.%, and 70 wt.%) were synthesized, termed PEEU40, PEEU50, PEEU60, and PEEU70, accordingly. The electrospinning method was used for the preparation of PEEU fiber meshes. The effects of PEEU fiber meshes with varying elasticities on the human umbilical vein endothelial cells (HUVECs) shape, growth, migration and angiogenic potential were characterized. To determine how the E-modulus of fiber meshes affects the osteogenic potential of hADSCs, the cellular and nuclear morphologies and osteogenic differentiation abilities were evaluated.
Results: With the increasing stiffness of PEEU fiber meshes, the aspect ratios of HUVECs cultivated on PEEU materials increased. HUVECs cultivated on high stiffness fiber meshes (4.5 ± 0.8 MPa) displayed a considerably greater proliferation rate and migratory velocity, in addition demonstrating increased tube formation capability, compared with those of the cells cultivated on lower stiffness fiber meshes (2.6 ± 0.8 MPa). Furthermore, in comparison to those cultivated on lower stiffness fiber meshes, hADSCs adhered to the highest stiffness fiber meshes PEEU70 had an elongated shape. The hADSCs grown on the softer PEEU40 fiber meshes showed a reduced nuclear aspect ratio (width to height) than those cultivated on the stiffer fiber meshes. Culturing hADSCs on stiffer fibers improved their osteogenic differentiation potential. Compared with cells cultured on PEEU40, osteocalcin expression and alkaline phosphatase (ALP) activity increased by 73 ± 10% and 43 ± 16%, respectively, in cells cultured on PEEU70.
Conclusion: The mechanical characteristics of the substrate are crucial in the modulation of cell behaviors. These findings indicate that adjusting the elasticity of fiber meshes might be a useful method for controlling the blood vessels development and regeneration. Furthermore, the mechanical characteristics of PEEU fiber meshes might be modified to control the osteogenic potential of hADSCs.
The self-employed faced strong income losses during the Covid-19 pandemic. Many governments introduced programs to financially support the self-employed during the pandemic, including Germany. The German Ministry for Economic Affairs announced a €50bn emergency-aid program in March 2020, offering one-off lump-sum payments of up to €15,000 to those facing substantial revenue declines. By reassuring the self- employed that the government ‘would not let them down’ during the crisis, the program had also the important aim of motivating the self-employed to get through the crisis. We investigate whether the program affected the confidence of the self-employed to survive the crisis using real-time online-survey data comprising more than 20,000 observations. We employ propensity score matching, making use of a rich set of variables that influence the subjective survival probability as main outcome measure. We observe that this program had significant effects, with the subjective survival probability of the self- employed being moderately increased. We reveal important effect heterogeneities with respect to education, industries, and speed of payment. Notably, positive effects only occur among those self-employed whose application was processed quickly. This suggests stress-induced waiting costs due to the uncertainty associated with the administrative processing and the overall pandemic situation. Our findings have policy implications for the design of support programs, while also contributing to the literature on the instruments and effects of entrepreneurship policy interventions in crisis situations.
The link between emotions and motor function has been known for decades but is still not clarified. The Adaptive Force (AF) describes the neuromuscular capability to adapt to increasing forces and was suggested to be especially vulnerable to interfering inputs. This study investigated the influence of pleasant an unpleasant food imagery on the manually assessed AF of elbow and hip flexors objectified by a handheld device in 12 healthy women. The maximal isometric AF was significantly reduced during unpleasant vs. pleasant imagery and baseline (p < 0.001, dz = 0.98–1.61). During unpleasant imagery, muscle lengthening started at 59.00 ± 22.50% of maximal AF, in contrast to baseline and pleasant imagery, during which the isometric position could be maintained mostly during the entire force increase up to ~97.90 ± 5.00% of maximal AF. Healthy participants showed an immediately impaired holding function triggered by unpleasant imagery, presumably related to negative emotions. Hence, AF seems to be suitable to test instantaneously the effect of emotions on motor function. Since musculoskeletal complaints can result from muscular instability, the findings provide insights into the understanding of the causal chain of linked musculoskeletal pain and mental stress. A case example (current stress vs. positive imagery) suggests that the approach presented in this study might have future implications for psychomotor diagnostics and therapeutics.
Entrepreneurial failure
(2022)
Although entrepreneurial failure (EF) is a fairly recent topic in entrepreneurship literature, the number of publications has been growing dynamically and particularly rapidly. Our systematic review maps and integrates the research on EF based on a multi-method approach to give structure and consistency to this fragmented field of research. The results reveal that the field revolves around six thematic clusters of EF: 1) Soft underpinnings of EF, 2) Contextuality of EF, 3) Perception of EF, 4) Two-sided effects of EF, 5) Multi-stage EF effects, and 6) Institutional drivers of EF. An integrative framework of the positive and negative effects of entrepreneurial failure is proposed, and a research agenda is suggested.
Entrepreneurship education (EE) has attracted much scholarly attention, showing exponential growth in publication and citation numbers. The research field has become broad, complex, and fragmented, making it increasingly difficult to oversee. Our research goal is to organise and integrate the previous literature. To this end, we use bibliometric analyses, differing from prior analyses, which are outdated or have a different focus. Our results show an immense growth in publications and citations over the last decade and an almost equal involvement of business and educational research. We identify the most productive and influential journals and authors. Our co-citation analysis reveals two research clusters, one focusing on psychological constructs relating to EE, and the other on entrepreneurial behaviour and new venture creation. Based on a review of the 25 most-cited articles on an annual basis, we identify and quantify the most relevant research themes and integrate them into a research framework that we propose for future research. A major finding is that extant research centres around the outcomes of entrepreneurship education, whereas its pedagogy is still mainly a black box.
Data profiling is the extraction of metadata from relational databases. An important class of metadata are multi-column dependencies. They come associated with two computational tasks. The detection problem is to decide whether a dependency of a given type and size holds in a database. The discovery problem instead asks to enumerate all valid dependencies of that type. We investigate the two problems for three types of dependencies: unique column combinations (UCCs), functional dependencies (FDs), and inclusion dependencies (INDs).
We first treat the parameterized complexity of the detection variants. We prove that the detection of UCCs and FDs, respectively, is W[2]-complete when parameterized by the size of the dependency. The detection of INDs is shown to be one of the first natural W[3]-complete problems. We further settle the enumeration complexity of the three discovery problems by presenting parsimonious equivalences with well-known enumeration problems. Namely, the discovery of UCCs is equivalent to the famous transversal hypergraph problem of enumerating the hitting sets of a hypergraph. The discovery of FDs is equivalent to the simultaneous enumeration of the hitting sets of multiple input hypergraphs. Finally, the discovery of INDs is shown to be equivalent to enumerating the satisfying assignments of antimonotone, 3-normalized Boolean formulas.
In the remainder of the thesis, we design and analyze discovery algorithms for unique column combinations. Since this is as hard as the general transversal hypergraph problem, it is an open question whether the UCCs of a database can be computed in output-polynomial time in the worst case. For the analysis, we therefore focus on instances that are structurally close to databases in practice, most notably, inputs that have small solutions. The equivalence between UCCs and hitting sets transfers the computational hardness, but also allows us to apply ideas from hypergraph theory to data profiling. We devise an discovery algorithm that runs in polynomial space on arbitrary inputs and achieves polynomial delay whenever the maximum size of any minimal UCC is bounded. Central to our approach is the extension problem for minimal hitting sets, that is, to decide for
a set of vertices whether they are contained in any minimal solution. We prove that this is yet another problem that is complete for the complexity class W[3], when parameterized by the size of the set that is to be extended. We also give several conditional lower bounds under popular hardness conjectures such as the Strong Exponential Time Hypothesis (SETH). The lower bounds suggest that the running time of our algorithm for the extension problem is close to optimal.
We further conduct an empirical analysis of our discovery algorithm on real-world databases to confirm that the hitting set perspective on data profiling has merits also in practice. We show that the resulting enumeration times undercut their theoretical worst-case bounds on practical data, and that the memory consumption of our method is much smaller than that of previous solutions. During the analysis we make two observations about the connection between databases and their corresponding hypergraphs. On the one hand, the hypergraph representations containing all relevant information are usually significantly smaller than the original inputs. On the other hand, obtaining those hypergraphs is the actual bottleneck of any practical application. The latter often takes much longer than enumerating the solutions, which is in stark contrast to the fact that the preprocessing is guaranteed to be polynomial while the enumeration may take exponential time.
To make the first observation rigorous, we introduce a maximum-entropy model for non-uniform random hypergraphs and prove that their expected number of minimal hyperedges undergoes a phase transition with respect to the total number of edges. The result also explains why larger databases may have smaller hypergraphs. Motivated by the second observation, we present a new kind of UCC discovery algorithm called Hitting Set Enumeration with Partial Information and Validation (HPIValid). It utilizes the fast enumeration times in practice in order to speed up the computation of the corresponding hypergraph. This way, we sidestep the bottleneck while maintaining the advantages of the hitting set perspective. An exhaustive empirical evaluation shows that HPIValid outperforms the current state of the art in UCC discovery. It is capable of processing databases that were previously out of reach for data profiling.
Long-term value creation is expected not only to be concerned with maximizing shareholder value but also includes the impact on other stakeholders and the environment. Environmental, social, and governance (ESG) issues are therefore gaining increasing importance, in line with the growing demand for corporate sustainability. ESG ratings foster the comparison of companies with respect to their sustainable practices. This study aims to investigate how ESG ratings impact financial performance in the European food industry. Ordinary least squares regression is applied to analyze the relation between ESG ratings and financial performance over a 4-year period from 2017 to 2020. The profitability measures Return on Assets (ROA) and Return on Equity (ROE) are employed as financial performance measures, while ESG ratings are obtained from the database CSRHub. Results show that higher ESG ratings are associated with better financial performance. Although the effect is modest in the present study, the findings support previous results that ESG ratings are positively related to financial performance. Nonetheless, they also highlight that ESG ratings strongly converge to the mean, which depicts the need to reassess whether ESG ratings are able to measure actual ESG behavior.
Essays in labor economics
(2022)
This thesis offers insights into the process of workers decisions to invest into work-related training. Specifically, the role of personality traits and attitudes is analysed. The aim is to understand whether such traits contribute to an under-investment into training. Importantly, general and specific training are distinguished, where the worker’s productivity increases in many firms in the former and only in the current firm in the latter case. Additionally, this thesis contributes to the evaluation of the German minimum wage introduction in 2015, identifying causal effects on wages and working hours.
Chapters two to four focus on the work-related training decision. First, individuals with an internal locus of control see a direct link between their own actions and their labor market success, while external individuals connect their outcomes to fate, luck, and other people. Consequently, it can be expected that internal individuals expect higher returns to training and are, thus, more willing to participate. The results reflect this hypothesis with internal individuals being more likely to participate in general (but not specific) training. Second, training can be viewed either as a risky investment or as an insurance against negative labor income shocks. In both cases, risk attitudes are expected to play a role in the decision process. The data point towards risk seeking individuals being more likely to participate in general (but not specific) training, and thus, training being viewed on average as a risky investment. Third, job satisfaction influences behavioral decisions in the job context, where dissatisfied workers may react by neglecting their duties, improving the situation or quitting the job. In the first case, dissatisfied workers are expected to invest less in training, while the latter two reactions could lead to higher participation rates amongst dissatisfied workers. The results suggest that on average dissatisfied workers are less likely to invest into training than satisfied workers. However, closer inspections of quit intentions and different sources of dissatisfaction paint less clear pictures, pointing towards the complexity of the job satisfaction construct.
Chapters five and six evaluate the introduction of the minimum wage in Germany in 2015. First, in 2015 an increase in the growth of hourly wages can be identified as a causal effect of the minimum wage introduction. However, at the same time, a reduction in the weekly working hours results in an overall unchanged growth in monthly earnings. When considering the effects in 2016, the decrease in weekly working hours disappears, resulting in a significant increase in the growth of monthly earnings due to the minimum wage. Importantly, the analysis suggests that the increase in hourly wages was not sufficient to ensure all workers receiving the minimum wage. This points to non-compliance being an issue in the first years after the minimum wage introduction.
Refugee youth in protracted humanitarian contexts are faced with limited access to quality education. They may sustain traumatic experiences from conflicts and discrimination yet have limited psychosocial support access. Comprehending the magnitude and effects of these challenges is vital for designing and executing educational interventions in such contexts. This study evaluates the implementation quality of the Youth Education Pack intervention through the lens of the Inter-agency Network for Education in Emergencies minimum standards framework. It explores the types of discrimination experienced by refugee youth in the Dadaab refugee camp in Kenya. Nine participants comprising refugee students (N = 2), former refugee students (N = 2), teachers (N = 3), and project supervisors (N = 2) participated in the study. The first author conducted interviews and observations in the camp. The data were qualitatively coded deductively and analysed in Nvivo 12. We found that the YEP intervention faced contextual challenges that hindered the achievement of the implementation quality standards outlined in the INEE minimum standards for education. Refugee youth and refugee teachers experienced various forms of discrimination, including at individual, institutional, and structural levels. We conclude that providing refugee youth with an inclusive and high-quality education is central to providing secure and long-term solutions to their challenges and adversities and may promote their psychosocial wellbeing.
Background
Non-invasive transcutaneous auricular vagus nerve stimulation (taVNS) has received tremendous attention as a potential neuromodulator of cognitive and affective functions, which likely exerts its effects via activation of the locus coeruleus-noradrenaline (LC-NA) system. Reliable effects of taVNS on markers of LC-NA system activity, however, have not been demonstrated yet.
Methods
The aim of the present study was to overcome previous limitations by pooling raw data from a large sample of ten taVNS studies (371 healthy participants) that collected salivary alpha-amylase (sAA) as a potential marker of central NA release.
Results
While a meta-analytic approach using summary statistics did not yield any significant effects, linear mixed model analyses showed that afferent stimulation of the vagus nerve via taVNS increased sAA levels compared to sham stimulation (b = 0.16, SE = 0.05, p = 0.001). When considering potential confounders of sAA, we further replicated previous findings on the diurnal trajectory of sAA activity.
Conclusion(s)
Vagal activation via taVNS increases sAA release compared to sham stimulation, which likely substantiates the assumption that taVNS triggers NA release. Moreover, our results highlight the benefits of data pooling and data sharing in order to allow stronger conclusions in research.
Plate tectonics describes the movement of rigid plates at the surface of the Earth as well as their complex deformation at three types of plate boundaries: 1) divergent boundaries such as rift zones and mid-ocean ridges, 2) strike-slip boundaries where plates grind past each other, such as the San Andreas Fault, and 3) convergent boundaries that form large mountain ranges like the Andes. The generally narrow deformation zones that bound the plates exhibit complex strain patterns that evolve through time. During this evolution, plate boundary deformation is driven by tectonic forces arising from Earth’s deep interior and from within the lithosphere, but also by surface processes, which erode topographic highs and deposit the resulting sediment into regions of low elevation. Through the combination of these factors, the surface of the Earth evolves in a highly dynamic way with several feedback mechanisms. At divergent boundaries, for example, tensional stresses thin the lithosphere, forcing uplift and subsequent erosion of rift flanks, which creates a sediment source. Meanwhile, the rift center subsides and becomes a topographic low where sediments accumulate. This mass transfer from foot- to hanging wall plays an important role during rifting, as it prolongs the activity of individual normal faults. When rifting continues, continents are eventually split apart, exhuming Earth’s mantle and creating new oceanic crust. Because of the complex interplay between deep tectonic forces that shape plate boundaries and mass redistribution at the Earth’s surface, it is vital to understand feedbacks between the two domains and how they shape our planet.
In this study I aim to provide insight on two primary questions: 1) How do divergent and strike-slip plate boundaries evolve? 2) How is this evolution, on a large temporal scale and a smaller structural scale, affected by the alteration of the surface through erosion and deposition? This is done in three chapters that examine the evolution of divergent and strike-slip plate boundaries using numerical models. Chapter 2 takes a detailed look at the evolution of rift systems using two-dimensional models. Specifically, I extract faults from a range of rift models and correlate them through time to examine how fault networks evolve in space and time. By implementing a two-way coupling between the geodynamic code ASPECT and landscape evolution code FastScape, I investigate how the fault network and rift evolution are influenced by the system’s erosional efficiency, which represents many factors like lithology or climate. In Chapter 3, I examine rift evolution from a three-dimensional perspective. In this chapter I study linkage modes for offset rifts to determine when fast-rotating plate-boundary structures known as continental microplates form. Chapter 4 uses the two-way numerical coupling between tectonics and landscape evolution to investigate how a strike-slip boundary responds to large sediment loads, and whether this is sufficient to form an entirely new type of flexural strike-slip basin.
The ongoing climate change is altering the living conditions for many organisms on this planet at an unprecedented pace. Hence, it is crucial for the survival of species to adapt to these changing conditions. In this dissertation Silene vulgaris is used as a model organism to understand the adaption strategies of widely distributed plant species to the current climate change. Especially plant species that possess a wide geographic range are expected to have a high phenotypic plasticity or to show genetic differentiation in response to the different climate conditions they grow in. However, they are often underrepresented in research.
In the greenhouse experiment presented in this thesis, I examined the phenotypic responses and plasticity in S. vulgaris to estimate its’ adaptation potential. Seeds from 25 wild European populations were collected along a latitudinal gradient and grown in a greenhouse under three different precipitation (65 mm, 75 mm, 90 mm) and two different temperature regimes (18°C, 21°C) that resembled a possible climate change scenario for central Europe. Afterwards different biomass and fecundity-related plant traits were measured.
The treatments significantly influenced the plants but did not reveal a latitudinal difference in response to climate treatments for most plant traits. The number of flowers per individual however, showed a stronger plasticity in northern European populations (e.g., Swedish populations) where numbers decreased more drastically with increased temperature and decreased precipitation.
To gain an even deeper understanding of the adaptation of S. vulgaris to climate change it is also important to reveal the underlying phylogeny of the sampled populations. Therefore, I analysed their population genetic structure through whole genome sequencing via ddRAD.
The sequencing revealed three major genetic clusters in the S. vulgaris populations sampled in Europe: one cluster comprised Southern European populations, one cluster Western European populations and another cluster contained central European populations. A following analysis of experimental trait responses among the clusters to the climate-change scenario showed that the genetic clusters significantly differed in biomass-related traits and in the days to flowering. However, half of the traits showed parallel response patterns to the experimental climate-change scenario.
In addition to the potential geographic and genetic adaptation differences to climate change this dissertation also deals with the response differences between the sexes in S. vulgaris. As a gynodioecious species populations of S. vulgaris consist of female and hermaphrodite
individuals and the sexes can differ in their morphological traits which is known as sexual dimorphism. As climate change is becoming an important factor influencing plant morphology it remains unclear if and how different sexes may respond in sexually dimorphic species. To examine this question the sex of each individual plant was determined during the greenhouse experiment and the measured plant traits were analysed accordingly. In general, hermaphrodites had a higher number of flowers but a lower number of leaves than females. With regards to the climate change treatment, I found that hermaphrodites showed a milder negative response to higher temperatures in the number of flowers produced and in specific leaf area (SLA) compared to females.
Synthesis – The significant treatment response in Silene vulgaris, independent of population origin in most traits suggests a high degree of universal phenotypic plasticity. Also, the three European intraspecific genetic lineages detected showed comparable parallel response patterns in half of the traits suggesting considerable phenotypic plasticity. Hence, plasticity might represent a possible adaptation strategy of this widely distributed species during ongoing and future climatic changes. The results on sexual dimorphism show that females and hermaphrodites are differing mainly in their number of flowers and females are affected more strongly by the experimental climate-change scenario. These results provide a solid knowledge basis on the sexual dimorphism in S. vulgaris under climate change, but further research is needed to determine the long-term impact on the breeding system for the species.
In summary this dissertation provides a comprehensive insight into the adaptation mechanisms and consequences of a widely distributed and gynodioecious plant species and leverages our understanding of the impact of anthropogenic climate change on plants.