Refine
Year of publication
- 2021 (2546) (remove)
Document Type
- Article (1574)
- Doctoral Thesis (275)
- Part of a Book (185)
- Postprint (157)
- Monograph/Edited Volume (101)
- Review (67)
- Conference Proceeding (43)
- Part of Periodical (36)
- Other (23)
- Working Paper (20)
- Master's Thesis (18)
- Habilitation Thesis (14)
- Report (11)
- Contribution to a Periodical (9)
- Bachelor Thesis (7)
- Course Material (4)
- Preprint (1)
- Sound (1)
Language
Keywords
- COVID-19 (23)
- machine learning (15)
- climate change (13)
- Migration (12)
- diffusion (12)
- Germany (11)
- embodied cognition (11)
- USA (10)
- exercise (10)
- Arabidopsis thaliana (9)
Institute
- Institut für Biochemie und Biologie (212)
- Institut für Physik und Astronomie (183)
- Institut für Geowissenschaften (169)
- Institut für Chemie (160)
- Bürgerliches Recht (128)
- Historisches Institut (120)
- Department Psychologie (103)
- Fachgruppe Betriebswirtschaftslehre (101)
- Fachgruppe Politik- & Verwaltungswissenschaft (100)
- Institut für Umweltwissenschaften und Geographie (100)
"BreaThink"
(2021)
Cognition is shaped by signals from outside and within the body. Following recent evidence of interoceptive signals modulating higher-level cognition, we examined whether breathing changes the production and perception of quantities. In Experiment 1, 22 adults verbally produced on average larger random numbers after inhaling than after exhaling. In Experiment 2, 24 further adults estimated the numerosity of dot patterns that were briefly shown after either inhaling or exhaling. Again, we obtained on average larger responses following inhalation than exhalation. These converging results extend models of situated cognition according to which higher-level cognition is sensitive to transient interoceptive states.
In den vergangenen Jahren hat sich die Politikdidaktik zunehmend mit dem Einsatz von Narrationen im Politikunterricht beschäftigt, denn neben Sachtexten bietet auch die Belletristik die Möglichkeit, sich mit politischen Themen auseinanderzusetzen. Insbesondere die Literatur von Ferdinand von Schirach hat in den letzten Jahren zunehmend Anklang in der Gesellschaft gefunden. Von Schirachs Texte greifen gesellschaftskritische Themen auf, beleuchten diese aus verschiedenen Perspektiven und fordern zur Meinungsbildung heraus. Aus diesem Grund weisen von Schirachs Narrationen ein hohes Potential für die Politische Bildung auf. Politische Bildung schließt auch die Rechterziehung ein. Der Fall Collini von Ferdinand von Schirach setzt sich sowohl mit rechtlichen, als auch mit politischen Themen im Sinne der Rechtserziehung auseinander. In der vorliegenden Masterarbeit wird der Frage nachgegangen, inwieweit der Roman Der Fall Collini von Ferdinand von Schirach als Narration eine Chance für politisch-rechtliches Lernen im Politikunterricht darstellt. Um die Forschungsfrage zu beantworten, werden die Lernchancen und -grenzen des Romans hinsichtlich seiner Thematik und seines Genres, sowie durch den Roman geförderten Kompetenzen herausgearbeitet und die durch ihn möglichen fächerübergreifenden Bezüge verdeutlicht. Durch die Auseinandersetzung mit von Schirachs Werk beschäftigen sich die Schülerinnen und Schüler mit politisch-rechtlichen Themen, wie dem Spannungsverhältnis von Recht und Gerechtigkeit, dem Ablauf von Strafgerichtsverfahren, dem theoretischen Anspruch des Rechtsstaates und dessen realen Schwächen. Zudem fördert die Auseinandersetzung mit dem Roman Der Fall Collini die vier fachbezogenen Kompetenzen der Politischen Bildung, sowie Multiperspektivität und exemplarisches Lernen. Des Weiteren verknüpft der Roman historische, politisch-rechtliche und moralisch-ethische Aspekte miteinander, wodurch fächerübergreifende Bezüge mit den Fächern Geschichte, Deutsch und L-E-R hergestellt werden können. Darüber hinaus spricht der Justizroman als Narration seine Leserinnen und Leser auch emotional an und fördert somit eine ganzheitliche und nachhaltige Wissensvermittlung im Sinne der Rechtserziehung. Es hat sich gezeigt, dass Der Fall Collini von Ferdinand von Schirach sich für die unterrichtliche Beschäftigung innerhalb der Politischen Bildung besonders eignet.
Ionic liquids are well known for their high gas absorption capacity. It is shown that this is not a solvent constant, but can be enhanced by another factor of 10 by pore confinement, here of the ionic liquid (IL) 1-ethyl-3-methylimidazolium acetate (EmimOAc) in the pores of carbon materials. A matrix of four different carbon compounds with micro- and mesopores as well as with and without nitrogen doping is utilized to investigate the influence of the carbons structure on the nitrogen uptake in the pore-confined EmimOAc. In general, the absorption is most improved for IL in micropores and in nitrogen-doped carbon. This effect is so large that it is already seen in TGA and DSC experiments. Due to the low vapor pressure of the IL, standard volumetric sorption experiments can be used to quantify details of this effect. It is reasoned that it is the change of the molecular arrangement of the ions in the restricted space of the pores that creates additional free volume to host molecular nitrogen.
Somewhat surprisingly, inverted ("O-down") CO adsorbates on NaCl(100) were recently observed experimentally after infrared vibrational excitation (Lau et al., Science, 2020, 367, 175-178). Here we characterize these species using periodic density functional theory and a quantum mechanical description of vibrations. We determine stationary points and minimum energy paths for CO inversion, for low (1/8 and 1/4 monolayers (ML)) and high (1 ML) coverages. Transition state theory is applied to estimate thermal rates for "C-down" to "O-down" isomerization and the reverse process. For the 1/4 ML p(1 x 1) structure, two-dimensional and three-dimensional potential energy surfaces and corresponding anharmonic vibrational eigenstates obtained from the time-independent nuclear Schrodinger equation are presented. We find (i) rather coverage-independent CO inversion energies (of about 0.08 eV or 8 kJ mol(-1) per CO) and corresponding classical activation energies for "C-down" to "O-down" isomerization (of about 0.15 eV or 14 kJ mol(-1) per CO); (ii) thermal isomerization rates at 22 K which are vanishingly small for the "C-down" to "O-down" isomerization but non-negligible for the back reaction; (iii) several "accidentally degenerate" pairs of eigenstates well below the barrier, each pair describing "C-down" to "O-down" localized states.
Being perceived as a foreigner regardless of one's generational status, citizenship, or self-identification is called foreigner objectification. This is a form of identity denial and is linked to psychological distress. To test how foreigner objectification could be measured in Europe, we assessed whether the Foreigner Objectification Scale demonstrated reliability and validity with German adolescents. The sample included 806 9th graders from 17 high schools. The results showed that the scale demonstrates good reliability, scalar measurement invariance across gender and citizenship status, and partial scalar measurement invariance across family heritage, generational status, and cultural self-identification. Adolescents who scored higher on the scale also reported greater school behavioral disengagement, lower life satisfaction, and stronger ethnic identity. Our findings suggest that the scale is psychometrically sound and is linked in theoretically consistent ways to adjustment and ethnic identity. We conclude that this scale offers another way to capture subtle discrimination experiences that add to a more comprehensive understanding of discrimination and the related implications in Europe.
The epitope imprinting approach applies exposed peptides as templates to synthesize Molecularly Imprinted Polymers (MIPs) for the recognition of the parent protein. While generally the template protein binding to such MIPs is considered to occur via the epitope-shaped cavities, unspecific interactions of the analyte with non-imprinted polymer as well as the detection method used may add to the complexity and interpretation of the target rebinding. To get new insights on the effects governing the rebinding of analytes, we electrosynthesized two epitope-imprinted polymers using the N-terminal pentapeptide VHLTP-amide of human hemoglobin (HbA) as the template. MIPs were prepared either by single-step electrosynthesis of scopoletin/pentapeptide mixtures or electropolymerization was performed after chemisorption of the cysteine extended VHLTP peptide. Rebinding of the target peptide and the parent HbA protein to the MIP nanofilms was quantified by square wave voltammetry using a redox probe gating, surface enhanced infrared absorption spectroscopy, and atomic force microscopy. While binding of the pentapeptide shows large influence of the amino acid sequence, all three methods revealed strong non-specific binding of HbA to both polyscopoletin-based MIPs with even higher affinities than the target peptides.
Participation has become an orthodoxy in the field of development, an essential element of projects and programmes. This book analyses participation in development interventions as an institutionalised expectation – a rationalized myth – and examines how organisations on different levels of government process it. At least two different objectives of participation are appropriate and legitimate for international organisations in the field: the empowerment of local beneficiaries and the achievement of programme goals. Both integrate participatory forums into the organisational logic of development interventions. Local administrations react to the institutionalised expectation with means-ends decoupling, where participatory forums are implemented superficially but de facto remain marginalised in local administrative processes and activities. The book furthermore provides a thick description of the organisationality of participation in development interventions. Participatory forums are shown to be a form of partial organisation. They establish an order in the relationship between administrations and citizens through the introduction of rules and the creation of a defined membership. At the same time, this order is found to be fragile and subject to criticism and negotiation.
Two decades ago, Tarana Burke started using the phrase ‘me too’ to release victims of sexual abuse and rape from their shame and to empower girls from minority communities. In 2017, actress Alyssa Milano made the hashtag #MeToo go viral. This article’s concern is with the role of testimonial practices in the context of sexual violence. While many feminists have claimed that the word of those who claim to being sexually violated by others (should) have political and/or epistemic priority, others have failed to recognize the harm and injury of instances of sexual violence that are not yet acknowledged as such and failed to listen to victims from marginalized social groups. In fact, some feminists have attacked #MeToo for mingling accounts of ‘proper’ sexual violence and accounts that are not ‘proper’ experiences of sexual violence. My aim in this article is to show why this critique is problematic and find a philosophically fruitful way to understand the #MeToo-movement as a movement that strives for moral and conceptual progress.
Seit dem Schuljahr 2020/21 gilt in Nordrhein-Westfalen ein neuer Kernlehrplan für die Realschule, Gesamtschule und Sekundarschule. Dafür haben wir gemeinsam mit Fachkräften aus dem Bundesland die #-Schulbuchreihen entwickelt.
Mit #Politik Wirtschaft – Nordrhein-Westfalen bieten wir Ihnen innovative und aktuelle Produkte für einen modernen Politik- und Wirtschaftsunterricht. Neben dem neuen Lehrplan sind die Vorgaben des Medienkompetenzrahmens und die besonderen Herausforderungen heterogener Lerngruppen berücksichtigt.
Wir bieten Ihnen einen problemorientierten und schülernahen Unterricht. Die Rubrik ”Gemeinsam aktiv“ ermöglicht ein selbstgesteuertes Lernen. Die Schülerinnen und Schüler erarbeiten sich projektartig größere Einheiten eines Kapitels. Sie können Ihren Unterricht einfach und schnell besonders vielfältig und spannend gestalten.
Durch Fallbeispiele werden die Schülerinnen und Schüler direkt angesprochen. Eine kreative Vielfalt aus Bild-, Grafik- und Textmaterial, aktivierende Aufgaben, Methoden-und Grundwissenseiten und ein Kompetenzcheck zum Abschluss der Großkapitel vervollständigen das Angebot.
Zu jeder Unterrichtseinheit wird passgenau zum Schulbuch unterschiedliches Differenzierungsmaterial (Texte in einfacher Sprache, Vorstrukturierung von Aufgaben u.v.m.) erstellt. Dieses steht Ihnen in unserem digitalen Lehrermaterial click & teach zur Verfügung und kann von Ihnen nach individuellen Bedürfnissen für einzelne digitale Schulbücher click & study freigeschaltet werden.
(Auf) Humboldts Spuren
(2021)
Vor seiner Besteigung des Antisana in Ecuador verbrachte Alexander von Humboldt mit seinem Expeditionsteam die Nacht vom 15. auf den 16. März 1802 in einer Hacienda am Fuße des Vulkangipfels, deren letztes bauliches Zeugnis eine steinerne Hütte darstellt. Bauforscherische Untersuchungen eines internationalen Forscherteams konnten die mehrschichtige Bau- und Reparaturgeschichte dieses Baudenkmals ermitteln und über eine Auswertung von Reiseberichten mehrerer Andenforscher die Nutzungsgeschichte des einzelnen Gebäudes und des gesamten Anwesens klären. Schließlich ergaben sich daraus neue Erkenntnisse zu Humboldts Aufenthalt am Antisana.
30 Jahre Universität Potsdam
(2021)
Das Referat für Presse- und Öffentlichkeitsarbeit hat anlässlich des 30-jährigen Universitätsjubiläums einen Bildband zusammengestellt. Auf 168 Seiten finden sich rund 400 Fotos aus der Geschichte der Potsdamer Alma Mater – von der Pressekonferenz, auf der Brandenburgs Wissenschaftsminister Dr. Hinrich Enderlein die Gründung der Universität bekannt gibt, bis zu den traurig-schönen menschenleeren Campusimpressionen Anfang 2021. Die Auswahl aus Zehntausenden Bildern, die im Laufe der Jahre entstanden sind, dokumentiert die rasante Entwicklung der drei Uni-Standorte Golm, Am Neuen Palais und Griebnitzsee, begleitet Studierende und Forschende verschiedener Generationen und zeigt Eindrücke von wichtigen großen und kleinen Veranstaltungen und Ereignissen sowie – vor allem – vom Alltag einer Universität.
Ground-penetrating radar (GPR) is a standard geophysical technique used to image near-surface structures in sedimentary environments. In such environments, GPR data acquisition and processing are increasingly following 3D strategies. However, the processed GPR data volumes are typically still interpreted using selected 2D slices and manual concepts such as GPR facies analyses. In seismic volume interpretation, the application of (semi-)automated and reproducible approaches such as 3D attribute analyses as well as the production of attribute-based facies models are common practices today. In contrast, the field of 3D GPR attribute analyses and corresponding facies models is largely untapped. We have developed and applied a workflow to produce 3D attribute-based GPR facies models comprising the dominant sedimentary reflection patterns in a GPR volume, which images complex sandy structures on the dune island of Spiekeroog (Northern Germany). After presenting our field site and details regarding our data acquisition and processing, we calculate and filter 3D texture attributes to generate a database comprising the dominant texture features of our GPR data. Then, we perform a dimensionality reduction of this database to obtain meta texture attributes, which we analyze and integrate using composite imaging and (also considering additional geometric information) fuzzy c-means cluster analysis resulting in a classified GPR facies model. Considering our facies model and a corresponding GPR facies chart, we interpret our GPR data set in terms of near-surface sedimentary units, the corresponding depositional environments, and the recent formation history at our field site. Thus, we demonstrate the potential of our workflow, which represents a novel and clear strategy to perform a more objective and consistent interpretation of 3D GPR data collected across different sedimentary environments.
Die Technologie des 3D-Drucks hat sich in den letzten Jahrzehnten rasant entwickelt. Im Industriebereich entstehen immer modernere und spezialisiertere Druckverfahren, im Hobby- und Privatanwenderbereich hingegen werden stetig kostengünstigere und einfacher zu bedienende Geräte zugänglich. Einzig im Bildungsbereich scheint das Themenfeld hingegen erst langsam eine Rolle zu spielen, obwohl sich zahlreiche Bezugspunkte für einen Einsatz in verschiedensten Fächern finden lassen. Insbesondere im Fach Wirtschaft-Arbeit-Technik sind die Schnittstellen zum Rahmenlehrplan Berlin/Brandenburg augenscheinlich, doch es liegen erst vereinzelt konkrete und systematische didaktische Konzepte und Vorschläge zur unterrichtspraktischen Einbettung vor. Die Verfasserin versucht daher in dieser Arbeit die Relevanz des Themas für die technische Bildung deutlich zu machen, eine kurze technische Einführung in das für einen schulischen Einsatz besonders geeignete FDM-Druckverfahren zu geben und daran anknüpfend konkrete Umsetzungsvorschläge aufzuzeigen: einerseits in Form eines allgemeinen Phasenmodells zur Planung von Technikunterricht sowie andererseits in Form eines exemplarischen Unterrichtskonzepts. Am Beispiel eines Schachsets wird verdeutlicht, wie Schülerinnen und Schüler zum Anfertigen der Konstruktionsunterlagen digitale CAD-Programme nutzen und anschließend mit Hilfe eines 3D-Druckers additiv fertigen können.
The prehistory of electrets is not known yet, but it is quite likely that the electrostatic charging behavior of amber (Greek: τò ηλεκτρoν, i.e., “electron”) already was familiar to people in ancient cultures (China, Egypt, Greece, etc.), before the Greek philosopher and scientist Thales of Miletus (6th century BCE)-or rather his disciples and followers-reported it in writing (cf. Figure 1). More than two millennia later, William Gilbert (1544–1603), the physician of Queen Elizabeth I, coined the term “electric” in his book De Magnete, Magneticisque Corporibus, et de Magno Magnete Tellure (1600) for dielectric materials that attract like amber and that included sulfur and glass [1]. The second half of the 18th century saw the invention of the electrophorus or electrophore [2], a capacitive electret device, in 1762 by Johan Carl Wilcke (1732–1796).
An easy-to-do synthesis for the hexanuclear niobium cluster compound [Nb6Cl12(CH3OH)(4)(OCH3)(2)] . DABCO . 1.66 CH2Cl2 has been developed. An one-pot reaction between the cluster precursor [Nb6Cl14(H2O)(4)] . 4H(2)O and methanol with the addition of DABCO leads to the crystallization of the title compound in high yield within a few minutes. The single-crystal X-ray structure of this cluster compound has been determined. Very strong, nearly symmetric intercluster hydrogen bonds Nb-6-MeO...H...OMe-Nb-6 are present between the cluster units. A bridging co-crystalline DABCO molecule is also involved in a three-dimensional hydrogen-bonding network.
In eye-movement control during reading, advanced process-oriented models have been developed to reproduce behavioral data. So far, model complexity and large numbers of model parameters prevented rigorous statistical inference and modeling of interindividual differences. Here we propose a Bayesian approach to both problems for one representative computational model of sentence reading (SWIFT; Engbert et al., Psychological Review, 112, 2005, pp. 777-813). We used experimental data from 36 subjects who read the text in a normal and one of four manipulated text layouts (e.g., mirrored and scrambled letters). The SWIFT model was fitted to subjects and experimental conditions individually to investigate between- subject variability. Based on posterior distributions of model parameters, fixation probabilities and durations are reliably recovered from simulated data and reproduced for withheld empirical data, at both the experimental condition and subject levels. A subsequent statistical analysis of model parameters across reading conditions generates model-driven explanations for observable effects between conditions.
Abstract
In recent years, feedforward neural networks (NNs) have been successfully applied to reconstruct global plasmasphere dynamics in the equatorial plane. These neural network‐based models capture the large‐scale dynamics of the plasmasphere, such as plume formation and erosion of the plasmasphere on the nightside. However, their performance depends strongly on the availability of training data. When the data coverage is limited or non‐existent, as occurs during geomagnetic storms, the performance of NNs significantly decreases, as networks inherently cannot learn from the limited number of examples. This limitation can be overcome by employing physics‐based modeling during strong geomagnetic storms. Physics‐based models show a stable performance during periods of disturbed geomagnetic activity if they are correctly initialized and configured. In this study, we illustrate how to combine the neural network‐ and physics‐based models of the plasmasphere in an optimal way by using data assimilation. The proposed approach utilizes advantages of both neural network‐ and physics‐based modeling and produces global plasma density reconstructions for both quiet and disturbed geomagnetic activity, including extreme geomagnetic storms. We validate the models quantitatively by comparing their output to the in‐situ density measurements from RBSP‐A for an 18‐month out‐of‐sample period from June 30, 2016 to January 01, 2018 and computing performance metrics. To validate the global density reconstructions qualitatively, we compare them to the IMAGE EUV images of the He+ particle distribution in the Earth's plasmasphere for a number of events in the past, including the Halloween storm in 2003.
Air pollution is a pressing issue that is associated with adverse effects on human health, ecosystems, and climate. Despite many years of effort to improve air quality, nitrogen dioxide (NO2) limit values are still regularly exceeded in Europe, particularly in cities and along streets. This study explores how concentrations of nitrogen oxides (NOx = NO + NO2) in European urban areas have changed over the last decades and how this relates to changes in emissions. To do so, the incremental approach was used, comparing urban increments (i.e. urban background minus rural concentrations) to total emissions, and roadside increments (i.e. urban roadside concentrations minus urban background concentrations) to traffic emissions. In total, nine European cities were assessed. The study revealed that potentially confounding factors like the impact of urban pollution at rural monitoring sites through atmospheric transport are generally negligible for NOx. The approach proves therefore particularly useful for this pollutant. The estimated urban increments all showed downward trends, and for the majority of the cities the trends aligned well with the total emissions. However, it was found that factors like a very densely populated surrounding or local emission sources in the rural area such as shipping traffic on inland waterways restrict the application of the approach for some cities. The roadside increments showed an overall very diverse picture in their absolute values and trends and also in their relation to traffic emissions. This variability and the discrepancies between roadside increments and emissions could be attributed to a combination of local influencing factors at the street level and different aspects introducing inaccuracies to the trends of the emis-sion inventories used, including deficient emission factors. Applying the incremental approach was evaluated as useful for long-term pan-European studies, but at the same time it was found to be restricted to certain regions and cities due to data availability issues. The results also highlight that using emission inventories for the prediction of future health impacts and compliance with limit values needs to consider the distinct variability in the concentrations not only across but also within cities.
Over the past decades, natural hazards, many of which are aggravated by climate change and reveal an increasing trend in frequency and intensity, have caused significant human and economic losses and pose a considerable obstacle to sustainable development. Hence, dedicated action toward disaster risk reduction is needed to understand the underlying drivers and create efficient risk mitigation plans. Such action is requested by the Sendai Framework for Disaster Risk Reduction 2015-2030 (SFDRR), a global agreement launched in 2015 that establishes stating priorities for action, e.g. an improved understanding of disaster risk. Turkey is one of the SFDRR contracting countries and has been severely affected by many natural hazards, in particular earthquakes and floods. However, disproportionately little is known about flood hazards and risks in Turkey. Therefore, this thesis aims to carry out a comprehensive analysis of flood hazards for the first time in Turkey from triggering drivers to impacts. It is intended to contribute to a better understanding of flood risks, improvements of flood risk mitigation and the facilitated monitoring of progress and achievements while implementing the SFDRR.
In order to investigate the occurrence and severity of flooding in comparison to other natural hazards in Turkey and provide an overview of the temporal and spatial distribution of flood losses, the Turkey Disaster Database (TABB) was examined for the years 1960-2014. The TABB database was reviewed through comparison with the Emergency Events Database (EM-DAT), the Dartmouth Flood Observatory database, the scientific literature and news archives. In addition, data on the most severe flood events between 1960 and 2014 were retrieved. These served as a basis for analyzing triggering mechanisms (i.e. atmospheric circulation and precipitation amounts) and aggravating pathways (i.e. topographic features, catchment size, land use types and soil properties). For this, a new approach was developed and the events were classified using hierarchical cluster analyses to identify the main influencing factor per event and provide additional information about the dominant flood pathways for severe floods. The main idea of the study was to start with the event impacts based on a bottom-up approach and identify the causes that created damaging events, instead of applying a model chain with long-term series as input and searching for potentially impacting events as model outcomes. However, within the frequency analysis of the flood-triggering circulation pattern types, it was discovered that events in terms of heavy precipitation were not included in the list of most severe floods, i.e. their impacts were not recorded in national and international loss databases but were mentioned in news archives and reported by the Turkish State Meteorological Service. This finding challenges bottom-up modelling approaches and underlines the urgent need for consistent event and loss documentation. Therefore, as a next step, the aim was to enhance the flood loss documentation by calibrating, validating and applying the United Nations Office for Disaster Risk Reduction (UNDRR) loss estimation method for the recent severe flood events (2015-2020). This provided, a consistent flood loss estimation model for Turkey, allowing governments to estimate losses as quickly as possible after events, e.g. to better coordinate financial aid.
This thesis reveals that, after earthquakes, floods have the second most destructive effects in Turkey in terms of human and economic impacts, with over 800 fatalities and US$ 885.7 million in economic losses between 1960 and 2020, and that more attention should be paid on the national scale. The clustering results of the dominant flood-producing mechanisms (e.g. circulation pattern types, extreme rainfall, sudden snowmelt) present crucial information regarding the source and pathway identification, which can be used as base information for hazard identification in the preliminary risk assessment process. The implementation of the UNDRR loss estimation model shows that the model with country-specific parameters, calibrated damage ratios and sufficient event documentation (i.e. physically damaged units) can be recommended in order to provide first estimates of the magnitude of direct economic losses, even shortly after events have occurred, since it performed well when estimates were compared to documented losses.
The presented results can contribute to improving the national disaster loss database in Turkey and thus enable a better monitoring of the national progress and achievements with regard to the targets stated by the SFDRR. In addition, the outcomes can be used to better characterize and classify flood events. Information on the main underlying factors and aggravating flood pathways further supports the selection of suitable risk reduction policies.
All input variables used in this thesis were obtained from publicly available data. The results are openly accessible and can be used for further research.
As an overall conclusion, it can be stated that consistent loss data collection and better event documentation should gain more attention for a reliable monitoring of the implementation of the SFDRR. Better event documentation should be established according to a globally accepted standard for disaster classification and loss estimation in Turkey. Ultimately, this enables stakeholders to create better risk mitigation actions based on clear hazard definitions, flood event classification and consistent loss estimations.
As AI technology is increasingly used in production systems, different approaches have emerged from highly decentralized small-scale AI at the edge level to centralized, cloud-based services used for higher-order optimizations. Each direction has disadvantages ranging from the lack of computational power at the edge level to the reliance on stable network connections with the centralized approach. Thus, a hybrid approach with centralized and decentralized components that possess specific abilities and interact is preferred. However, the distribution of AI capabilities leads to problems in self-adapting learning systems, as knowledgebases can diverge when no central coordination is present. Edge components will specialize in distinctive patterns (overlearn), which hampers their adaptability for different cases. Therefore, this paper aims to present a concept for a distributed interchangeable knowledge base in CPPS. The approach is based on various AI components and concepts for each participating node. A service-oriented infrastructure allows a decentralized, loosely coupled architecture of the CPPS. By exchanging knowledge bases between nodes, the overall system should become more adaptive, as each node can “forget” their present specialization.
Exendin-4 is a pharmaceutical peptide used in the control of insulin secretion. Structural information on exendin-4 and related peptides especially on the level of quaternary structure is scarce. We present the first published association equilibria of exendin-4 directly measured by static and dynamic light scattering. We show that exendin-4 oligomerization is pH dependent and that these oligomers are of low compactness. We relate our experimental results to a structural hypothesis to describe molecular details of exendin-4 oligomers. Discussion of the validity of this hypothesis is based on NMR, circular dichroism and fluorescence spectroscopy, and light scattering data on exendin-4 and a set of exendin-4 derived peptides. The essential forces driving oligomerization of exendin-4 are helix–helix interactions and interactions of a conserved hydrophobic moiety. Our structural hypothesis suggests that key interactions of exendin-4 monomers in the experimentally supported trimer take place between a defined helical segment and a hydrophobic triangle constituted by the Phe22 residues of the three monomeric subunits. Our data rationalize that Val19 might function as an anchor in the N-terminus of the interacting helix-region and that Trp25 is partially shielded in the oligomer by C-terminal amino acids of the same monomer. Our structural hypothesis suggests that the Trp25 residues do not interact with each other, but with C-terminal Pro residues of their own monomers.
Exendin-4 is a pharmaceutical peptide used in the control of insulin secretion. Structural information on exendin-4 and related peptides especially on the level of quaternary structure is scarce. We present the first published association equilibria of exendin-4 directly measured by static and dynamic light scattering. We show that exendin-4 oligomerization is pH dependent and that these oligomers are of low compactness. We relate our experimental results to a structural hypothesis to describe molecular details of exendin-4 oligomers. Discussion of the validity of this hypothesis is based on NMR, circular dichroism and fluorescence spectroscopy, and light scattering data on exendin-4 and a set of exendin-4 derived peptides. The essential forces driving oligomerization of exendin-4 are helix–helix interactions and interactions of a conserved hydrophobic moiety. Our structural hypothesis suggests that key interactions of exendin-4 monomers in the experimentally supported trimer take place between a defined helical segment and a hydrophobic triangle constituted by the Phe22 residues of the three monomeric subunits. Our data rationalize that Val19 might function as an anchor in the N-terminus of the interacting helix-region and that Trp25 is partially shielded in the oligomer by C-terminal amino acids of the same monomer. Our structural hypothesis suggests that the Trp25 residues do not interact with each other, but with C-terminal Pro residues of their own monomers.
Transition path theory (TPT) for diffusion processes is a framework for analyzing the transitions of multiscale ergodic diffusion processes between disjoint metastable subsets of state space. Most methods for applying TPT involve the construction of a Markov state model on a discretization of state space that approximates the underlying diffusion process. However, the assumption of Markovianity is difficult to verify in practice, and there are to date no known error bounds or convergence results for these methods. We propose a Monte Carlo method for approximating the forward committor, probability current, and streamlines from TPT for diffusion processes. Our method uses only sample trajectory data and partitions of state space based on Voronoi tessellations. It does not require the construction of a Markovian approximating process. We rigorously prove error bounds for the approximate TPT objects and use these bounds to show convergence to their exact counterparts in the limit of arbitrarily fine discretization. We illustrate some features of our method by application to a process that solves the Smoluchowski equation on a triple-well potential.
The manuscript describes the phytochemical investigation of the roots, leaves and stem bark of Millettia lasiantha resulting in the isolation of twelve compounds including two new isomeric isoflavones lascoumestan and las-coumaronochromone. The structures of the new compounds were determined using different spectroscopic techniques.
Ground subsidence caused by natural or anthropogenic processes affects major urban areas worldwide. Sinkhole formation and infrastructure fractures have intensified in the federal capital of Maceio (Alagoas, Brazil) since early 2018, forcing authorities to relocate affected residents and place buildings under demolition. In this study, we present a 16-year history (2004-2020) of surface displacement, which shows precursory deformations in 2004-2005, reaching a maximum cumulative subsidence of approximately 200 cm near the Mundau Lagoon coast in November 2020. By integrating the displacement observations with numerical source modelling, we suggest that extensive subsidence can be primarily associated with the removal of localized, deep-seated material at the location and depth where salt is mined. We discuss the accelerating subsidence rates, influence of severe precipitation events on the aforementioned geological instability, and related hazards. This study suggests that feedback destabilization mechanisms may arise in evaporite systems due to anthropogenic activities, fostering enhanced and complex superficial ground deformation.
A drop of immunity
(2021)
A expulsão do Éden
(2021)
A temática da migração está intimamente vinculada à história humana, desde a narrativa bíblica da expulsão do paraíso. O ser humano não apenas empregou técnicas cada vez mais sofisticadas para a violência, como também transmitiu, através dos séculos, técnicas de conservação e uso de seu saberconviver. Nesse sentido móvel da história, e em consonância com as literaturas do mundo, a partir de suas diversas origens, é possível dizer que existe um “Homo migrans” desde que existe o “Homo sapiens”. Assim, é possível afirmar que as ideias territoriais ou territorializantes com proveniência histórico-espacial permitem, vez ou outra, reconhecer seus esforços para filtrar e isolar a dimensão histórico-móvel e vetorial da história como narrativa, para tentar construir, com a ajuda de ideias estáticas, novos lugares da promessa ou da perda, da abundância ou da queda.
We present a new numerical algorithm to solve the recently derived equations of two-moment cosmic ray hydrodynamics (CRHD). The algorithm is implemented as a module in the moving mesh AREPO code. Therein, the anisotropic transport of cosmic rays (CRs) along magnetic field lines is discretized using a path-conservative finite volume method on the unstructured time-dependent Voronoi mesh of AREPO. The interaction of CRs and gyroresonant Alfven waves is described by short time-scale source terms in the CRHD equations. We employ a custom-made semi-implicit adaptive time stepping source term integrator to accurately integrate this interaction on the small light-crossing time of the anisotropic transport step. Both the transport and the source term integration step are separated from the evolution of the magnetohydrodynamical equations using an operator split approach. The new algorithm is tested with a variety of test problems, including shock tubes, a perpendicular magnetized discontinuity, the hydrodynamic response to a CR overpressure, CR acceleration of a warm cloud, and a CR blast wave, which demonstrate that the coupling between CR and magnetohydrodynamics is robust and accurate. We demonstrate the numerical convergence of the presented scheme using new linear and non-linear analytic solutions.
Image feature detection is a key task in computer vision. Scale Invariant Feature Transform (SIFT) is a prevalent and well known algorithm for robust feature detection. However, it is computationally demanding and software implementations are not applicable for real-time performance. In this paper, a versatile and pipelined hardware implementation is proposed, that is capable of computing keypoints and rotation invariant descriptors on-chip. All computations are performed in single precision floating-point format which makes it possible to implement the original algorithm with little alteration. Various rotation resolutions and filter kernel sizes are supported for images of any resolution up to ultra-high definition. For full high definition images, 84 fps can be processed. Ultra high definition images can be processed at 21 fps.
A grammar of authority?
(2021)
Directive Speech Acts (dsas) are a major feature of historical pragmatics, specifically in research on historical (im)politeness. However, for Classical French, there is a lack of research on related phenomena. In our contribution, we present two recently constructed corpora covering the period of Classical French, sermo and apwcf. We present these corpora in terms of their genre characteristics on a communicative-functional and socio-pragmatic level. Based on the observation that, both in sermo and apwcf, dsas frequently occur together with terms of address, we analyse and manually code a sample based on this co-occurrence, and we compare the results with regard to special features in the individual corpora. The emerging patterns show a clear correspondence between socio-pragmatic factors and the linguistic means used to realise dsas. We propose that these results can be interpreted as signs of an underlying "grammar of authority".
We prove a homology vanishing theorem for graphs with positive Bakry-' Emery curvature, analogous to a classic result of Bochner on manifolds [3]. Specifically, we prove that if a graph has positive curvature at every vertex, then its first homology group is trivial, where the notion of homology that we use for graphs is the path homology developed by Grigor'yan, Lin, Muranov, and Yau [11]. We moreover prove that the fundamental group is finite for graphs with positive Bakry-' Emery curvature, analogous to a classic result of Myers on manifolds [22]. The proofs draw on several separate areas of graph theory, including graph coverings, gain graphs, and cycle spaces, in addition to the Bakry-Emery curvature, path homology, and graph homotopy. The main results follow as a consequence of several different relationships developed among these different areas. Specifically, we show that a graph with positive curvature cannot have a non-trivial infinite cover preserving 3-cycles and 4-cycles, and give a combinatorial interpretation of the first path homology in terms of the cycle space of a graph. Furthermore, we relate gain graphs to graph homotopy and the fundamental group developed by Grigor'yan, Lin, Muranov, and Yau [12], and obtain an alternative proof of their result that the abelianization of the fundamental group of a graph is isomorphic to the first path homology over the integers.
We present a new autoclave that enables in situ characterization of hydrothermal fluids at high pressures and high temperatures at synchrotron x-ray radiation sources. The autoclave has been specifically designed to enable x-ray absorption spectroscopy in fluids with applications to mineral solubility and element speciation analysis in hydrothermal fluids in complex compositions. However, other applications, such as Raman spectroscopy, in high-pressure fluids are also possible with the autoclave. First experiments were run at pressures between 100 and 600 bars and at temperatures between 25 degrees C and 550 degrees C, and preliminary results on scheelite dissolution in fluids of different compositions show that the autoclave is well suited to study the behavior of ore-forming metals at P-T conditions relevant to the Earth's crust.
Context. The spectroscopic class of subdwarf A-type (sdA) stars has come into focus in recent years because of their possible link to extremely low-mass white dwarfs, a rare class of objects resulting from binary evolution. Although most sdA stars are consistent with metal-poor halo main-sequence stars, the formation and evolution of a fraction of these stars are still matters of debate. Aims. The identification of photometric variability can help to put further constraints on the evolutionary status of sdA stars, in particular through the analysis of pulsations. Moreover, the binary ratio, which can be deduced from eclipsing binaries and ellipsoidal variables, is important as input for stellar models. In order to search for variability due to either binarity or pulsations in objects of the spectroscopic sdA class, we have extracted all available high precision light curves from the Kepler K2 mission.
Methods. We have performed a thorough time series analysis on all available light curves, employing three different methods. Frequencies with a signal-to-noise ratio higher than four have been used for further analysis.
Results. From the 25 targets, 13 turned out to be variables of different kinds (i.e., classical pulsating stars, ellipsoidal and cataclysmic variables, eclipsing binaries, and rotationally induced variables). For the remaining 12 objects, a variability threshold was determined.
A few months before his death, A. v. Humboldt attended the celebration in honor of the 127th birthday of George Washington at the US legation in Berlin. A letter to the American Envoy, Joseph A. Wright (1810 – 1867), underlines Humboldt’s admiration for the fi rst president of the United States. At the same time Humboldt asked the diplomat to mail a letter to the German-American Bernard Moses (1832 – 1897) in Clinton, Louisiana, who had named his son Alexander Humboldt Moses (grave on the Hebrew Rest Cemetery #2 in New Orleans, burial plot A, 12, 5). It appears to be possible that the Moses family still owns Humboldt’s letter.
We introduce a logic-based incremental approach to graph repair, generating a sound and complete (upon termination) overview of least-changing graph repairs from which a user may select a graph repair based on non-formalized further requirements. This incremental approach features delta preservation as it allows to restrict the generation of graph repairs to delta-preserving graph repairs, which do not revert the additions and deletions of the most recent consistency-violating graph update. We specify consistency of graphs using the logic of nested graph conditions, which is equivalent to first-order logic on graphs. Technically, the incremental approach encodes if and how the graph under repair satisfies a graph condition using the novel data structure of satisfaction trees, which are adapted incrementally according to the graph updates applied. In addition to the incremental approach, we also present two state-based graph repair algorithms, which restore consistency of a graph independent of the most recent graph update and which generate additional graph repairs using a global perspective on the graph under repair. We evaluate the developed algorithms using our prototypical implementation in the tool AutoGraph and illustrate our incremental approach using a case study from the graph database domain.
Aging in speech production is a multidimensional process. Biological, cognitive, social, and communicative factors can change over time, stay relatively stable, or may even compensate for each other. In this longitudinal work, we focus on stability and change at the laryngeal and supralaryngeal levels in the discourse particle euh produced by 10 older French-speaking females at two times, 10 years apart. Recognizing the multiple discourse roles of euh, we divided out occurrences according to utterance position. We quantified the frequency of euh, and evaluated acoustic changes in formants, fundamental frequency, and voice quality across time and utterance position. Results showed that euh frequency was stable with age. The only acoustic measure that revealed an age effect was harmonics-to-noise ratio, showing less noise at older ages. Other measures mostly varied with utterance position, sometimes in interaction with age. Some voice quality changes could reflect laryngeal adjustments that provide for airflow conservation utterance-finally. The data suggest that aging effects may be evident in some prosodic positions (e.g., utterance-final position), but not others (utterance-initial position). Thus, it is essential to consider the interactions among these factors in future work and not assume that vocal aging is evident throughout the signal.
A matter of concern
(2021)
Neurons are post-mitotic cells in the brain and their integrity is of central importance to avoid neurodegeneration. Yet, the inability of self-replenishment of post-mitotic cells results in the need to withstand challenges from numerous stressors during life. Neurons are exposed to oxidative stress due to high oxygen consumption during metabolic activity in the brain. Accordingly, DNA damage can occur and accumulate, resulting in genome instability. In this context, imbalances in brain trace element homeostasis are a matter of concern, especially regarding iron, copper, manganese, zinc, and selenium. Although trace elements are essential for brain physiology, excess and deficient conditions are considered to impair neuronal maintenance. Besides increasing oxidative stress, DNA damage response and repair of oxidative DNA damage are affected by trace elements. Hence, a balanced trace element homeostasis is of particular importance to safeguard neuronal genome integrity and prevent neuronal loss. This review summarises the current state of knowledge on the impact of deficient, as well as excessive iron, copper, manganese, zinc, and selenium levels on neuronal genome stability
The large literature that aims to find evidence of climate migration delivers mixed findings. This meta-regression analysis i) summarizes direct links between adverse climatic events and migration, ii) maps patterns of climate migration, and iii) explains the variation in outcomes. Using a set of limited dependent variable models, we meta-analyze thus-far the most comprehensive sample of 3,625 estimates from 116 original studies and produce novel insights on climate migration. We find that extremely high temperatures and drying conditions increase migration. We do not find a significant effect of sudden-onset events. Climate migration is most likely to emerge due to contemporaneous events, to originate in rural areas and to take place in middle-income countries, internally, to cities. The likelihood to become trapped in affected areas is higher for women and in low-income countries, particularly in Africa. We uniquely quantify how pitfalls typical for the broader empirical climate impact literature affect climate migration findings. We also find evidence of different publication biases.
With the downscaling of CMOS technologies, the radiation-induced Single Event Transient (SET) effects in combinational logic have become a critical reliability issue for modern integrated circuits (ICs) intended for operation under harsh radiation conditions. The SET pulses generated in combinational logic may propagate through the circuit and eventually result in soft errors. It has thus become an imperative to address the SET effects in the early phases of the radiation-hard IC design. In general, the soft error mitigation solutions should accommodate both static and dynamic measures to ensure the optimal utilization of available resources. An efficient soft-error-aware design should address synergistically three main aspects: (i) characterization and modeling of soft errors, (ii) multi-level soft error mitigation, and (iii) online soft error monitoring. Although significant results have been achieved, the effectiveness of SET characterization methods, accuracy of predictive SET models, and efficiency of SET mitigation measures are still critical issues. Therefore, this work addresses the following topics: (i) Characterization and modeling of SET effects in standard combinational cells, (ii) Static mitigation of SET effects in standard combinational cells, and (iii) Online particle detection, as a support for dynamic soft error mitigation.
Since the standard digital libraries are widely used in the design of radiation-hard ICs, the characterization of SET effects in standard cells and the availability of accurate SET models for the Soft Error Rate (SER) evaluation are the main prerequisites for efficient radiation-hard design. This work introduces an approach for the SPICE-based standard cell characterization with the reduced number of simulations, improved SET models and optimized SET sensitivity database. It has been shown that the inherent similarities in the SET response of logic cells for different input levels can be utilized to reduce the number of required simulations. Based on characterization results, the fitting models for the SET sensitivity metrics (critical charge, generated SET pulse width and propagated SET pulse width) have been developed. The proposed models are based on the principle of superposition, and they express explicitly the dependence of the SET sensitivity of individual combinational cells on design, operating and irradiation parameters. In contrast to the state-of-the-art characterization methodologies which employ extensive look-up tables (LUTs) for storing the simulation results, this work proposes the use of LUTs for storing the fitting coefficients of the SET sensitivity models derived from the characterization results. In that way the amount of characterization data in the SET sensitivity database is reduced significantly.
The initial step in enhancing the robustness of combinational logic is the application of gate-level mitigation techniques. As a result, significant improvement of the overall SER can be achieved with minimum area, delay and power overheads. For the SET mitigation in standard cells, it is essential to employ the techniques that do not require modifying the cell structure. This work introduces the use of decoupling cells for improving the robustness of standard combinational cells. By insertion of two decoupling cells at the output of a target cell, the critical charge of the cell’s output node is increased and the attenuation of short SETs is enhanced. In comparison to the most common gate-level techniques (gate upsizing and gate duplication), the proposed approach provides better SET filtering. However, as there is no single gate-level mitigation technique with optimal performance, a combination of multiple techniques is required. This work introduces a comprehensive characterization of gate-level mitigation techniques aimed to quantify their impact on the SET robustness improvement, as well as introduced area, delay and power overhead per gate. By characterizing the gate-level mitigation techniques together with the standard cells, the required effort in subsequent SER analysis of a target design can be reduced. The characterization database of the hardened standard cells can be utilized as a guideline for selection of the most appropriate mitigation solution for a given design.
As a support for dynamic soft error mitigation techniques, it is important to enable the online detection of energetic particles causing the soft errors. This allows activating the power-greedy fault-tolerant configurations based on N-modular redundancy only at the high radiation levels. To enable such a functionality, it is necessary to monitor both the particle flux and the variation of particle LET, as these two parameters contribute significantly to the system SER. In this work, a particle detection approach based on custom-sized pulse stretching inverters is proposed. Employing the pulse stretching inverters connected in parallel enables to measure the particle flux in terms of the number of detected SETs, while the particle LET variations can be estimated from the distribution of SET pulse widths. This approach requires a purely digital processing logic, in contrast to the standard detectors which require complex mixed-signal processing. Besides the possibility of LET monitoring, additional advantages of the proposed particle detector are low detection latency and power consumption, and immunity to error accumulation.
The results achieved in this thesis can serve as a basis for establishment of an overall soft-error-aware database for a given digital library, and a comprehensive multi-level radiation-hard design flow that can be implemented with the standard IC design tools. The following step will be to evaluate the achieved results with the irradiation experiments.
Atmospheric water vapour content is a key variable that controls the development of deep convective storms and rainfall extremes over the central Andes. Direct measurements of water vapour are challenging; however, recent developments in microwave processing allow the use of phase delays from L-band radar to measure the water vapour content throughout the atmosphere: Global Navigation Satellite System (GNSS)-based integrated water vapour (IWV) monitoring shows promising results to measure vertically integrated water vapour at high temporal resolutions. Previous works also identified convective available potential energy (CAPE) as a key climatic variable for the formation of deep convective storms and rainfall in the central Andes. Our analysis relies on GNSS data from the Argentine Continuous Satellite Monitoring Network, Red Argentina de Monitoreo Satelital Continuo (RAMSAC) network from 1999 to 2013. CAPE is derived from version 2.0 of the ECMWF’s (European Centre for Medium-Range Weather Forecasts) Re-Analysis (ERA-interim) and rainfall from the TRMM (Tropical Rainfall Measuring Mission) product. In this study, we first analyse the rainfall characteristics of two GNSS-IWV stations by comparing their complementary cumulative distribution function (CCDF). Second, we separately derive the relation between rainfall vs. CAPE and GNSS-IWV. Based on our distribution fitting analysis, we observe an exponential relation of rainfall to GNSS-IWV. In contrast, we report a power-law relationship between the daily mean value of rainfall and CAPE at the GNSS-IWV station locations in the eastern central Andes that is close to the theoretical relationship based on parcel theory. Third, we generate a joint regression model through a multivariable regression analysis using CAPE and GNSS-IWV to explain the contribution of both variables in the presence of each other to extreme rainfall during the austral summer season. We found that rainfall can be characterised with a higher statistical significance for higher rainfall quantiles, e.g., the 0.9 quantile based on goodness-of-fit criterion for quantile regression. We observed different contributions of CAPE and GNSS-IWV to rainfall for each station for the 0.9 quantile. Fourth, we identify the temporal relation between extreme rainfall (the 90th, 95th, and 99th percentiles) and both GNSS-IWV and CAPE at 6 h time steps. We observed an increase before the rainfall event and at the time of peak rainfall—both for GNSS-integrated water vapour and CAPE. We show higher values of CAPE and GNSS-IWV for higher rainfall percentiles (99th and 95th percentiles) compared to the 90th percentile at a 6-h temporal scale. Based on our correlation analyses and the dynamics of the time series, we show that both GNSS-IWV and CAPE had comparable magnitudes, and we argue to consider both climatic variables when investigating their effect on rainfall extremes.
Atmospheric water vapour content is a key variable that controls the development of deep convective storms and rainfall extremes over the central Andes. Direct measurements of water vapour are challenging; however, recent developments in microwave processing allow the use of phase delays from L-band radar to measure the water vapour content throughout the atmosphere: Global Navigation Satellite System (GNSS)-based integrated water vapour (IWV) monitoring shows promising results to measure vertically integrated water vapour at high temporal resolutions. Previous works also identified convective available potential energy (CAPE) as a key climatic variable for the formation of deep convective storms and rainfall in the central Andes. Our analysis relies on GNSS data from the Argentine Continuous Satellite Monitoring Network, Red Argentina de Monitoreo Satelital Continuo (RAMSAC) network from 1999 to 2013. CAPE is derived from version 2.0 of the ECMWF’s (European Centre for Medium-Range Weather Forecasts) Re-Analysis (ERA-interim) and rainfall from the TRMM (Tropical Rainfall Measuring Mission) product. In this study, we first analyse the rainfall characteristics of two GNSS-IWV stations by comparing their complementary cumulative distribution function (CCDF). Second, we separately derive the relation between rainfall vs. CAPE and GNSS-IWV. Based on our distribution fitting analysis, we observe an exponential relation of rainfall to GNSS-IWV. In contrast, we report a power-law relationship between the daily mean value of rainfall and CAPE at the GNSS-IWV station locations in the eastern central Andes that is close to the theoretical relationship based on parcel theory. Third, we generate a joint regression model through a multivariable regression analysis using CAPE and GNSS-IWV to explain the contribution of both variables in the presence of each other to extreme rainfall during the austral summer season. We found that rainfall can be characterised with a higher statistical significance for higher rainfall quantiles, e.g., the 0.9 quantile based on goodness-of-fit criterion for quantile regression. We observed different contributions of CAPE and GNSS-IWV to rainfall for each station for the 0.9 quantile. Fourth, we identify the temporal relation between extreme rainfall (the 90th, 95th, and 99th percentiles) and both GNSS-IWV and CAPE at 6 h time steps. We observed an increase before the rainfall event and at the time of peak rainfall—both for GNSS-integrated water vapour and CAPE. We show higher values of CAPE and GNSS-IWV for higher rainfall percentiles (99th and 95th percentiles) compared to the 90th percentile at a 6-h temporal scale. Based on our correlation analyses and the dynamics of the time series, we show that both GNSS-IWV and CAPE had comparable magnitudes, and we argue to consider both climatic variables when investigating their effect on rainfall extremes.
We consider a sequential cascade of molecular first-reaction events towards a terminal reaction centre in which each reaction step is controlled by diffusive motion of the particles. The model studied here represents a typical reaction setting encountered in diverse molecular biology systems, in which, e.g. a signal transduction proceeds via a series of consecutive 'messengers': the first messenger has to find its respective immobile target site triggering a launch of the second messenger, the second messenger seeks its own target site and provokes a launch of the third messenger and so on, resembling a relay race in human competitions. For such a molecular relay race taking place in infinite one-, two- and three-dimensional systems, we find exact expressions for the probability density function of the time instant of the terminal reaction event, conditioned on preceding successful reaction events on an ordered array of target sites. The obtained expressions pertain to the most general conditions: number of intermediate stages and the corresponding diffusion coefficients, the sizes of the target sites, the distances between them, as well as their reactivities are arbitrary.
We consider a sequential cascade of molecular first-reaction events towards a terminal reaction centre in which each reaction step is controlled by diffusive motion of the particles. The model studied here represents a typical reaction setting encountered in diverse molecular biology systems, in which, e.g. a signal transduction proceeds via a series of consecutive 'messengers': the first messenger has to find its respective immobile target site triggering a launch of the second messenger, the second messenger seeks its own target site and provokes a launch of the third messenger and so on, resembling a relay race in human competitions. For such a molecular relay race taking place in infinite one-, two- and three-dimensional systems, we find exact expressions for the probability density function of the time instant of the terminal reaction event, conditioned on preceding successful reaction events on an ordered array of target sites. The obtained expressions pertain to the most general conditions: number of intermediate stages and the corresponding diffusion coefficients, the sizes of the target sites, the distances between them, as well as their reactivities are arbitrary.
In an attempt to pave the way for more extensive Computer Science Education (CSE) coverage in K-12, this research developed and made a preliminary evaluation of a blended-learning Introduction to CS program based on an academic MOOC. Using an academic MOOC that is pedagogically effective and engaging, such a program may provide teachers with disciplinary scaffolds and allow them to focus their attention on enhancing students’ learning experience and nurturing critical 21st-century skills such as self-regulated learning. As we demonstrate, this enabled us to introduce an academic level course to middle-school students. In this research, we developed the principals and initial version of such a program, targeting ninth-graders in science-track classes who learn CS as part of their standard curriculum. We found that the middle-schoolers who participated in the program achieved academic results on par with undergraduate students taking this MOOC for academic credit. Participating students also developed a more accurate perception of the essence of CS as a scientific discipline. The unplanned school closure due to the COVID19 pandemic outbreak challenged the research but underlined the advantages of such a MOOCbased blended learning program above classic pedagogy in times of global or local crises that lead to school closure. While most of the science track classes seem to stop learning CS almost entirely, and the end-of-year MoE exam was discarded, the program’s classes smoothly moved to remote learning mode, and students continued to study at a pace similar to that experienced before the school shut down.
Despite advanced seismological techniques, automatic source characterization for microseismic earthquakes remains difficult and challenging since current inversion and modelling of high-frequency signals are complex and time consuming. For real-time applications such as induced seismicity monitoring, the application of standard methods is often not fast enough for true complete real-time information on seismic sources. In this paper, we present an alternative approach based on recent advances in deep learning for rapid source-parameter estimation of microseismic earthquakes. The seismic inversion is represented in compact form by two convolutional neural networks, with individual feature extraction, and a fully connected neural network, for feature aggregation, to simultaneously obtain full moment tensor and spatial location of microseismic sources. Specifically, a multibranch neural network algorithm is trained to encapsulate the information about the relationship between seismic waveforms and underlying point-source mechanisms and locations. The learning-based model allows rapid inversion (within a fraction of second) once input data are available. A key advantage of the algorithm is that it can be trained using synthetic seismic data only, so it is directly applicable to scenarios where there are insufficient real data for training. Moreover, we find that the method is robust with respect to perturbations such as observational noise and data incompleteness (missing stations). We apply the new approach on synthesized and example recorded small magnitude (M <= 1.6) earthquakes at the Hellisheioi geothermal field in the Hengill area, Iceland. For the examined events, the model achieves excellent performance and shows very good agreement with the inverted solutions determined through standard methodology. In this study, we seek to demonstrate that this approach is viable for microseismicity real-time estimation of source parameters and can be integrated into advanced decision-support tools for controlling induced seismicity.