Refine
Year of publication
- 2022 (33) (remove)
Document Type
- Article (27)
- Doctoral Thesis (5)
- Postprint (1)
Language
- English (33)
Is part of the Bibliography
- yes (33)
Keywords
- graphs (2)
- knots (2)
- molecular weaving (2)
- networks (2)
- periodic entanglement (2)
- tangles (2)
- Animal movement modeling (1)
- Arctic haze (1)
- Bayesian inference (1)
- Bayesian inversion (1)
Institute
- Institut für Mathematik (33) (remove)
Extreme value statistics is a popular and frequently used tool to model the occurrence of large earthquakes. The problem of poor statistics arising from rare events is addressed by taking advantage of the validity of general statistical properties in asymptotic regimes. In this note, I argue that the use of extreme value statistics for the purpose of practically modeling the tail of the frequency-magnitude distribution of earthquakes can produce biased and thus misleading results because it is unknown to what degree the tail of the true distribution is sampled by data. Using synthetic data allows to quantify this bias in detail. The implicit assumption that the true M-max is close to the maximum observed magnitude M-max,M-observed restricts the class of the potential models a priori to those with M-max = M-max,M-observed + Delta M with an increment Delta M approximate to 0.5... 1.2. This corresponds to the simple heuristic method suggested by Wheeler (2009) and labeled :M-max equals M-obs plus an increment." The incomplete consideration of the entire model family for the frequency-magnitude distribution neglects, however, the scenario of a large so far unobserved earthquake.
Ground motion with strong-velocity pulses can cause significant damage to buildings and structures at certain periods; hence, knowing the period and velocity amplitude of such pulses is critical for earthquake structural engineering.
However, the physical factors relating the scaling of pulse periods with magnitude are poorly understood.
In this study, we investigate moderate but damaging earthquakes (M-w 6-7) and characterize ground- motion pulses using the method of Shahi and Baker (2014) while considering the potential static-offset effects.
We confirm that the within-event variability of the pulses is large. The identified pulses in this study are mostly from strike-slip-like earthquakes. We further perform simulations using the freq uency-wavenumber algorithm to investigate the causes of the variability of the pulse periods within and between events for moderate strike-slip earthquakes.
We test the effect of fault dips, and the impact of the asperity locations and sizes. The simulations reveal that the asperity properties have a high impact on the pulse periods and amplitudes at nearby stations.
Our results emphasize the importance of asperity characteristics, in addition to earthquake magnitudes for the occurrence and properties of pulses produced by the forward directivity effect.
We finally quantify and discuss within- and between-event variabilities of pulse properties at short distances.
We propose a global geomagnetic field model for the last 14 thousand years, based on thermoremanent records. We call the model ArchKalmag14k. ArchKalmag14k is constructed by modifying recently proposed algorithms, based on space-time correlations. Due to the amount of data and complexity of the model, the full Bayesian posterior is numerically intractable. To tackle this, we sequentialize the inversion by implementing a Kalman-filter with a fixed time step. Every step consists of a prediction, based on a degree dependent temporal covariance, and a correction via Gaussian process regression. Dating errors are treated via a noisy input formulation. Cross correlations are reintroduced by a smoothing algorithm and model parameters are inferred from the data. Due to the specific statistical nature of the proposed algorithms, the model comes with space and time-dependent uncertainty estimates. The new model ArchKalmag14k shows less variation in the large-scale degrees than comparable models. Local predictions represent the underlying data and agree with comparable models, if the location is sampled well. Uncertainties are bigger for earlier times and in regions of sparse data coverage. We also use ArchKalmag14k to analyze the appearance and evolution of the South Atlantic anomaly together with reverse flux patches at the core-mantle boundary, considering the model uncertainties. While we find good agreement with earlier models for recent times, our model suggests a different evolution of intensity minima prior to 1650 CE. In general, our results suggest that prior to 6000 BCE the data is not sufficient to support global models.
The geomagnetic main field is vital for live on Earth, as it shields our habitat against the solar wind and cosmic rays. It is generated by the geodynamo in the Earth’s outer core and has a rich dynamic on various timescales. Global models of the field are used to study the interaction of the field and incoming charged particles, but also to infer core dynamics and to feed numerical simulations of the geodynamo. Modern satellite missions, such as the SWARM or the CHAMP mission, support high resolution reconstructions of the global field. From the 19 th century on, a global network of magnetic observatories has been established. It is growing ever since and global models can be constructed from the data it provides. Geomagnetic field models that extend further back in time rely on indirect observations of the field, i.e. thermoremanent records such as burnt clay or volcanic rocks and sediment records from lakes and seas. These indirect records come with (partially very large) uncertainties, introduced by the complex measurement methods and the dating procedure.
Focusing on thermoremanent records only, the aim of this thesis is the development of a new modeling strategy for the global geomagnetic field during the Holocene, which takes the uncertainties into account and produces realistic estimates of the reliability of the model. This aim is approached by first considering snapshot models, in order to address the irregular spatial distribution of the records and the non-linear relation of the indirect observations to the field itself. In a Bayesian setting, a modeling algorithm based on Gaussian process regression is developed and applied to binned data. The modeling algorithm is then extended to the temporal domain and expanded to incorporate dating uncertainties. Finally, the algorithm is sequentialized to deal with numerical challenges arising from the size of the Holocene dataset.
The central result of this thesis, including all of the aspects mentioned, is a new global geomagnetic field model. It covers the whole Holocene, back until 12000 BCE, and we call it ArchKalmag14k. When considering the uncertainties that are produced together with the model, it is evident that before 6000 BCE the thermoremanent database is not sufficient to support global models. For times more recent, ArchKalmag14k can be used to analyze features of the field under consideration of posterior uncertainties. The algorithm for generating ArchKalmag14k can be applied to different datasets and is provided to the community as an open source python package.
Hidden semi-Markov models generalise hidden Markov models by explicitly modelling the time spent in a given state, the so-called dwell time, using some distribution defined on the natural numbers. While the (shifted) Poisson and negative binomial distribution provide natural choices for such distributions, in practice, parametric distributions can lack the flexibility to adequately model the dwell times. To overcome this problem, a penalised maximum likelihood approach is proposed that allows for a flexible and data-driven estimation of the dwell-time distributions without the need to make any distributional assumption. This approach is suitable for direct modelling purposes or as an exploratory tool to investigate the latent state dynamics. The feasibility and potential of the suggested approach is illustrated in a simulation study and by modelling muskox movements in northeast Greenland using GPS tracking data. The proposed method is implemented in the R-package PHSMM which is available on CRAN.
Model uncertainty quantification is an essential component of effective data assimilation. Model errors associated with sub-grid scale processes are often represented through stochastic parameterizations of the unresolved process. Many existing Stochastic Parameterization schemes are only applicable when knowledge of the true sub-grid scale process or full observations of the coarse scale process are available, which is typically not the case in real applications. We present a methodology for estimating the statistics of sub-grid scale processes for the more realistic case that only partial observations of the coarse scale process are available. Model error realizations are estimated over a training period by minimizing their conditional sum of squared deviations given some informative covariates (e.g., state of the system), constrained by available observations and assuming that the observation errors are smaller than the model errors. From these realizations a conditional probability distribution of additive model errors given these covariates is obtained, allowing for complex non-Gaussian error structures. Random draws from this density are then used in actual ensemble data assimilation experiments. We demonstrate the efficacy of the approach through numerical experiments with the multi-scale Lorenz 96 system using both small and large time scale separations between slow (coarse scale) and fast (fine scale) variables. The resulting error estimates and forecasts obtained with this new method are superior to those from two existing methods.
Background
Cytochrome P450 (CYP) 3A contributes to the metabolism of many approved drugs. CYP3A perpetrator drugs can profoundly alter the exposure of CYP3A substrates. However, effects of such drug-drug interactions are usually reported as maximum effects rather than studied as time-dependent processes. Identification of the time course of CYP3A modulation can provide insight into when significant changes to CYP3A activity occurs, help better design drug-drug interaction studies, and manage drug-drug interactions in clinical practice.
Objective
We aimed to quantify the time course and extent of the in vivo modulation of different CYP3A perpetrator drugs on hepatic CYP3A activity and distinguish different modulatory mechanisms by their time of onset, using pharmacologically inactive intravenous microgram doses of the CYP3A-specific substrate midazolam, as a marker of CYP3A activity.
Methods
Twenty-four healthy individuals received an intravenous midazolam bolus followed by a continuous infusion for 10 or 36 h. Individuals were randomized into four arms: within each arm, two individuals served as a placebo control and, 2 h after start of the midazolam infusion, four individuals received the CYP3A perpetrator drug: voriconazole (inhibitor, orally or intravenously), rifampicin (inducer, orally), or efavirenz (activator, orally). After midazolam bolus administration, blood samples were taken every hour (rifampicin arm) or every 15 min (remaining study arms) until the end of midazolam infusion. A total of 1858 concentrations were equally divided between midazolam and its metabolite, 1'-hydroxymidazolam. A nonlinear mixed-effects population pharmacokinetic model of both compounds was developed using NONMEM (R). CYP3A activity modulation was quantified over time, as the relative change of midazolam clearance encountered by the perpetrator drug, compared to the corresponding clearance value in the placebo arm.
Results
Time course of CYP3A modulation and magnitude of maximum effect were identified for each perpetrator drug. While efavirenz CYP3A activation was relatively fast and short, reaching a maximum after approximately 2-3 h, the induction effect of rifampicin could only be observed after 22 h, with a maximum after approximately 28-30 h followed by a steep drop to almost baseline within 1-2 h. In contrast, the inhibitory impact of both oral and intravenous voriconazole was prolonged with a steady inhibition of CYP3A activity followed by a gradual increase in the inhibitory effect until the end of sampling at 8 h. Relative maximum clearance changes were +59.1%, +46.7%, -70.6%, and -61.1% for efavirenz, rifampicin, oral voriconazole, and intravenous voriconazole, respectively.
Conclusions
We could distinguish between different mechanisms of CYP3A modulation by the time of onset. Identification of the time at which clearance significantly changes, per perpetrator drug, can guide the design of an optimal sampling schedule for future drug-drug interaction studies. The impact of a short-term combination of different perpetrator drugs on the paradigm CYP3A substrate midazolam was characterized and can define combination intervals in which no relevant interaction is to be expected.
The motivation for this work was the question of reliability and robustness of seismic tomography. The problem is that many earth models exist which can describe the underlying ground motion records equally well. Most algorithms for reconstructing earth models provide a solution, but rarely quantify their variability. If there is no way to verify the imaged structures, an interpretation is hardly reliable. The initial idea was to explore the space of equivalent earth models using Bayesian inference. However, it quickly became apparent that the rigorous quantification of tomographic uncertainties could not be accomplished within the scope of a dissertation.
In order to maintain the fundamental concept of statistical inference, less complex problems from the geosciences are treated instead. This dissertation aims to anchor Bayesian inference more deeply in the geosciences and to transfer knowledge from applied mathematics. The underlying idea is to use well-known methods and techniques from statistics to quantify the uncertainties of inverse problems in the geosciences. This work is divided into three parts:
Part I introduces the necessary mathematics and should be understood as a kind of toolbox. With a physical application in mind, this section provides a compact summary of all methods and techniques used. The introduction of Bayesian inference makes the beginning. Then, as a special case, the focus is on regression with Gaussian processes under linear transformations. The chapters on the derivation of covariance functions and the approximation of non-linearities are discussed in more detail.
Part II presents two proof of concept studies in the field of seismology. The aim is to present the conceptual application of the introduced methods and techniques with moderate complexity. The example about traveltime tomography applies the approximation of non-linear relationships. The derivation of a covariance function using the wave equation is shown in the example of a damped vibrating string. With these two synthetic applications, a consistent concept for the quantification of modeling uncertainties has been developed.
Part III presents the reconstruction of the Earth's archeomagnetic field. This application uses the whole toolbox presented in Part I and is correspondingly complex. The modeling of the past 1000 years is based on real data and reliably quantifies the spatial modeling uncertainties. The statistical model presented is widely used and is under active development.
The three applications mentioned are intentionally kept flexible to allow transferability to similar problems. The entire work focuses on the non-uniqueness of inverse problems in the geosciences. It is intended to be of relevance to those interested in the concepts of Bayesian inference.
Variational bayesian inference for nonlinear hawkes process with gaussian process self-effects
(2022)
Traditionally, Hawkes processes are used to model time-continuous point processes with history dependence. Here, we propose an extended model where the self-effects are of both excitatory and inhibitory types and follow a Gaussian Process. Whereas previous work either relies on a less flexible parameterization of the model, or requires a large amount of data, our formulation allows for both a flexible model and learning when data are scarce. We continue the line of work of Bayesian inference for Hawkes processes, and derive an inference algorithm by performing inference on an aggregated sum of Gaussian Processes. Approximate Bayesian inference is achieved via data augmentation, and we describe a mean-field variational inference approach to learn the model parameters. To demonstrate the flexibility of the model we apply our methodology on data from different domains and compare it to previously reported results.
Model-informed precision dosing (MIPD) is a quantitative dosing framework that combines prior knowledge on the drug-disease-patient system with patient data from therapeutic drug/ biomarker monitoring (TDM) to support individualized dosing in ongoing treatment. Structural models and prior parameter distributions used in MIPD approaches typically build on prior clinical trials that involve only a limited number of patients selected according to some exclusion/inclusion criteria. Compared to the prior clinical trial population, the patient population in clinical practice can be expected to also include altered behavior and/or increased interindividual variability, the extent of which, however, is typically unknown. Here, we address the question of how to adapt and refine models on the level of the model parameters to better reflect this real-world diversity. We propose an approach for continued learning across patients during MIPD using a sequential hierarchical Bayesian framework. The approach builds on two stages to separate the update of the individual patient parameters from updating the population parameters. Consequently, it enables continued learning across hospitals or study centers, because only summary patient data (on the level of model parameters) need to be shared, but no individual TDM data. We illustrate this continued learning approach with neutrophil-guided dosing of paclitaxel. The present study constitutes an important step toward building confidence in MIPD and eventually establishing MIPD increasingly in everyday therapeutic use.
We construct and examine the prototype of a deep learning-based ground-motion model (GMM) that is both fully data driven and nonergodic. We formulate ground-motion modeling as an image processing task, in which a specific type of neural network, the U-Net, relates continuous, horizontal maps of earthquake predictive parameters to sparse observations of a ground-motion intensity measure (IM). The processing of map-shaped data allows the natural incorporation of absolute earthquake source and observation site coordinates, and is, therefore, well suited to include site-, source-, and path-specific amplification effects in a nonergodic GMM. Data-driven interpolation of the IM between observation points is an inherent feature of the U-Net and requires no a priori assumptions. We evaluate our model using both a synthetic dataset and a subset of observations from the KiK-net strong motion network in the Kanto basin in Japan. We find that the U-Net model is capable of learning the magnitude???distance scaling, as well as site-, source-, and path-specific amplification effects from a strong motion dataset. The interpolation scheme is evaluated using a fivefold cross validation and is found to provide on average unbiased predictions. The magnitude???distance scaling as well as the site amplification of response spectral acceleration at a period of 1 s obtained for the Kanto basin are comparable to previous regional studies.
Randomised one-step time integration methods for deterministic operator differential equations
(2022)
Uncertainty quantification plays an important role in problems that involve inferring a parameter of an initial value problem from observations of the solution. Conrad et al. (Stat Comput 27(4):1065-1082, 2017) proposed randomisation of deterministic time integration methods as a strategy for quantifying uncertainty due to the unknown time discretisation error. We consider this strategy for systems that are described by deterministic, possibly time-dependent operator differential equations defined on a Banach space or a Gelfand triple. Our main results are strong error bounds on the random trajectories measured in Orlicz norms, proven under a weaker assumption on the local truncation error of the underlying deterministic time integration method. Our analysis establishes the theoretical validity of randomised time integration for differential equations in infinite-dimensional settings.
According to Radzikowski’s celebrated results, bisolutions of a wave operator on a globally hyperbolic spacetime are of the Hadamard form iff they are given by a linear combination of distinguished parametrices i2(G˜aF−G˜F+G˜A−G˜R) in the sense of Duistermaat and Hörmander [Acta Math. 128, 183–269 (1972)] and Radzikowski [Commun. Math. Phys. 179, 529 (1996)]. Inspired by the construction of the corresponding advanced and retarded Green operator GA, GR as done by Bär, Ginoux, and Pfäffle {Wave Equations on Lorentzian Manifolds and Quantization [European Mathematical Society (EMS), Zürich, 2007]}, we construct the remaining two Green operators GF, GaF locally in terms of Hadamard series. Afterward, we provide the global construction of i2(G˜aF−G˜F), which relies on new techniques such as a well-posed Cauchy problem for bisolutions and a patching argument using Čech cohomology. This leads to global bisolutions of the Hadamard form, each of which can be chosen to be a Hadamard two-point-function, i.e., the smooth part can be adapted such that, additionally, the symmetry and the positivity condition are exactly satisfied.
We present a technique for the enumeration of all isotopically distinct ways of tiling a hyperbolic surface of finite genus, possibly nonorientable and with punctures and boundary. This generalizes the enumeration using Delaney--Dress combinatorial tiling theory of combinatorial classes of tilings to isotopy classes of tilings. To accomplish this, we derive an action of the mapping class group of the orbifold associated to the symmetry group of a tiling on the set of tilings. We explicitly give descriptions and presentations of semipure mapping class groups and of tilings as decorations on orbifolds. We apply this enumerative result to generate an array of isotopically distinct tilings of the hyperbolic plane with symmetries generated by rotations that are commensurate with the threedimensional symmetries of the primitive, diamond, and gyroid triply periodic minimal surfaces, which have relevance to a variety of physical systems.
In this paper, we examine conditioning of the discretization of the Helmholtz problem. Although the discrete Helmholtz problem has been studied from different perspectives, to the best of our knowledge, there is no conditioning analysis for it. We aim to fill this gap in the literature. We propose a novel method in 1D to observe the near-zero eigenvalues of a symmetric indefinite matrix. Standard classification of ill-conditioning based on the matrix condition number is not true for the discrete Helmholtz problem. We relate the ill-conditioning of the discretization of the Helmholtz problem with the condition number of the matrix. We carry out analytical conditioning analysis in 1D and extend our observations to 2D with numerical observations. We examine several discretizations. We find different regions in which the condition number of the problem shows different characteristics. We also explain the general behavior of the solutions in these regions.
Alpine ecosystems on the Tibetan Plateau are being threatened by ongoing climate warming and intensified human activities. Ecological time-series obtained from sedimentary ancient DNA (sedaDNA) are essential for understanding past ecosystem and biodiversity dynamics on the Tibetan Plateau and their responses to climate change at a high taxonomic resolution. Hitherto only few but promising studies have been published on this topic. The potential and limitations of using sedaDNA on the Tibetan Plateau are not fully understood. Here, we (i) provide updated knowledge of and a brief introduction to the suitable archives, region-specific taphonomy, state-of-the-art methodologies, and research questions of sedaDNA on the Tibetan Plateau; (ii) review published and ongoing sedaDNA studies from the Tibetan Plateau; and (iii) give some recommendations for future sedaDNA study designs. Based on the current knowledge of taphonomy, we infer that deep glacial lakes with freshwater and high clay sediment input, such as those from the southern and southeastern Tibetan Plateau, may have a high potential for sedaDNA studies. Metabarcoding (for microorganisms and plants), metagenomics (for ecosystems), and hybridization capture (for prehistoric humans) are three primary sedaDNA approaches which have been successfully applied on the Tibetan Plateau, but their power is still limited by several technical issues, such as PCR bias and incompleteness of taxonomic reference databases. Setting up high-quality and open-access regional taxonomic reference databases for the Tibetan Plateau should be given priority in the future. To conclude, the archival, taphonomic, and methodological conditions of the Tibetan Plateau are favorable for performing sedaDNA studies. More research should be encouraged to address questions about long-term ecological dynamics at ecosystem scale and to bring the paleoecology of the Tibetan Plateau into a new era.
Conventional embeddings of the edge-graphs of Platonic polyhedra, {f,z}, where f,z denote the number of edges in each face and the edge-valence at each vertex, respectively, are untangled in that they can be placed on a sphere (S-2) such that distinct edges do not intersect, analogous to unknotted loops, which allow crossing-free drawings of S-1 on the sphere. The most symmetric (flag-transitive) realizations of those polyhedral graphs are those of the classical Platonic polyhedra, whose symmetries are *2fz, according to Conway's two-dimensional (2D) orbifold notation (equivalent to Schonflies symbols I-h, O-h, and T-d). Tangled Platonic {f,z} polyhedra-which cannot lie on the sphere without edge-crossings-are constructed as windings of helices with three, five, seven,... strands on multigenus surfaces formed by tubifying the edges of conventional Platonic polyhedra, have (chiral) symmetries 2fz (I, O, and T), whose vertices, edges, and faces are symmetrically identical, realized with two flags. The analysis extends to the "theta(z)" polyhedra, {2,z}. The vertices of these symmetric tangled polyhedra overlap with those of the Platonic polyhedra; however, their helicity requires curvilinear (or kinked) edges in all but one case. We show that these 2fz polyhedral tangles are maximally symmetric; more symmetric embeddings are necessarily untangled. On one hand, their topologies are very constrained: They are either self-entangled graphs (analogous to knots) or mutually catenated entangled compound polyhedra (analogous to links). On the other hand, an endless variety of entanglements can be realized for each topology. Simpler examples resemble patterns observed in synthetic organometallic materials and clathrin coats in vivo.
An explicit Dobrushin uniqueness region for Gibbs point processes with repulsive interactions
(2022)
We present a uniqueness result for Gibbs point processes with interactions that come from a non-negative pair potential; in particular, we provide an explicit uniqueness region in terms of activity z and inverse temperature beta. The technique used relies on applying to the continuous setting the classical Dobrushin criterion. We also present a comparison to the two other uniqueness methods of cluster expansion and disagreement percolation, which can also be applied for this type of interaction.
We discuss Neumann problems for self-adjoint Laplacians on (possibly infinite) graphs. Under the assumption that the heat semigroup is ultracontractive we discuss the unique solvability for non-empty subgraphs with respect to the vertex boundary and provide analytic and probabilistic representations for Neumann solutions. A second result deals with Neumann problems on canonically compactifiable graphs with respect to the Royden boundary and provides conditions for unique solvability and analytic and probabilistic representations.