Refine
Has Fulltext
- no (117)
Year of publication
- 2017 (117) (remove)
Document Type
- Other (117) (remove)
Is part of the Bibliography
- yes (117)
Keywords
- Internet (2)
- MOOC (2)
- affect (2)
- carbon dioxide (2)
- embodied cognition (2)
- 2.5D Treemaps (1)
- Absorption kinetics (1)
- Aluminium (1)
- Aluminium adjuvants (1)
- Aufklarung (1)
- BMI (1)
- Bandwidth (1)
- Buntsandstein (1)
- CO2 storage monitoring (1)
- Cloud Native Applications (1)
- Cloud-Security (1)
- DPP-4 inhibitors (1)
- Dialysis patients (1)
- Digital Learning Factory (1)
- Dipeptidyl peptidase IV (1)
- Durkheim (1)
- Durkheim’s German Reception, Max Weber, Georg Simmel, Jürgen Habermas (1)
- E-Learning (1)
- ERPs (1)
- Educational Technology (1)
- Enlightenment (1)
- Exploratory interfaces (1)
- Gamification (1)
- Geoeletrical imaging (1)
- Growth faltering (1)
- Growth modelling (1)
- In vitro dissolution (1)
- Industrial IoT Competences (1)
- Information Visualization (1)
- Massive Open Online Courses (1)
- Media retrieval (1)
- Memory management (1)
- Mobile Learning (1)
- Mobiles (1)
- Mortality (1)
- Multi-perspective Views (1)
- Multidimensional scaling (1)
- Obesity (1)
- Offline-Enabled (1)
- Overview plus Detail (1)
- P300 (1)
- PHEMA (1)
- Parallel programming (1)
- Performance analysis (1)
- Photon Density Wave Spectroscopy (1)
- SAW impedance sensor (1)
- Secular trend (1)
- Security-as-a-Service (1)
- Serum intact-parathyroid hormone level (1)
- Strategic growth adjustment (1)
- Student Training (1)
- Toxicokinetic modelling (1)
- Treemaps (1)
- Ubiquitous (1)
- User study (1)
- Vocational Training (1)
- Vulnerability Assessment (1)
- abstract concepts (1)
- action words (1)
- acute kidney injury (1)
- age of acquisition (1)
- agreement processing (1)
- algae cultivation (1)
- anaphor resolution (1)
- bilingualism (1)
- blind feeling (1)
- brain rhythms (1)
- brain synchronization (1)
- carbon cycle (1)
- child development (1)
- climate change (1)
- coefficient of determination (1)
- complex networks (1)
- connectivity (1)
- contingent encounters (1)
- critical period for language (1)
- damage (1)
- data based model (1)
- digital rock physics (1)
- digital technologies (1)
- dissolved (1)
- e-learning (1)
- effective elastic properties (1)
- eighteenth century (1)
- emotions (1)
- encoding (1)
- enjoyment (1)
- epilepsy (1)
- exercise (1)
- feelings (1)
- fermentation (1)
- fiber spectroscopy (1)
- filler-gap dependencies (1)
- gas storage (1)
- geothermal reservoir (1)
- gliptins (1)
- graph analysis (1)
- historiography (1)
- human rights (1)
- hybrid nanomaterials (1)
- hydrogen (1)
- implicit (1)
- interference (1)
- ischemia reperfusion injury (1)
- language (1)
- language acquisition (1)
- learning (1)
- lifespan (1)
- mattering (1)
- memory (1)
- memory retrieval (1)
- mental simulation (1)
- meteorological extremes (1)
- methane (1)
- microfluidic (1)
- mobile phone (1)
- modernity (1)
- multiple light scattering (1)
- multivariate regression (1)
- neural synchonization (1)
- non-linear dynamics (1)
- norepinephrine (1)
- numerical (1)
- numerical simulation (1)
- old/new effect (1)
- permanent downhole electrode array (1)
- physical activity (1)
- process analytical technology (1)
- regulation (1)
- renewable energy (1)
- risk analysis (1)
- saline aquifer (1)
- scenario analysis (1)
- sensation (1)
- sensitive periods (1)
- sentence comprehension (1)
- smartphones (1)
- storage capacity (1)
- tablet computers (1)
- transcutaneous vagus nerve stimulation (1)
- tropical cyclones (1)
- vulnerability (1)
- water management (1)
- wondering (1)
Institute
- Institut für Biochemie und Biologie (22)
- Institut für Physik und Astronomie (17)
- Institut für Geowissenschaften (11)
- Department Sport- und Gesundheitswissenschaften (10)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (10)
- Department Psychologie (8)
- Institut für Ernährungswissenschaft (6)
- Department Linguistik (5)
- Institut für Informatik und Computational Science (5)
- Institut für Chemie (4)
Recently, Kocyan & Wiland-Szymańska (2016) have published a thorough research article on one of the outstanding members of the family Hypoxidaceae on the Seychelles, which resulted in the raise of a new genus (Friedmannia Kocyan & Wiland-Szymańska 2016: 60) to accommodate the former Curculigo seychellensis Bojer ex Baker (1877: 368). However, it has turned out that the name Friedmannia Chantanachat & Bold (1962: 45) already exists in literature for a green alga, which renders the new hypoxid genus illegitimate (Melbourne Code; McNeill et al. 2012). Therefore, we assign a new generic epithet to Curculigo seychellensis.
Selection of initial points, the number of clusters and finding proper clusters centers are still the main challenge in clustering processes. In this paper, we suggest genetic algorithm based method which searches several solution spaces simultaneously. The solution spaces are population groups consisting of elements with similar structure. Elements in a group have the same size, while elements in different groups are of different sizes. The proposed algorithm processes the population in groups of chromosomes with one gene, two genes to k genes. These genes hold corresponding information about the cluster centers. In the proposed method, the crossover and mutation operators can accept parents with different sizes; this can lead to versatility in population and information transfer among sub-populations. We implemented the proposed method and evaluated its performance against some random datasets and the Ruspini dataset as well. The experimental results show that the proposed method could effectively determine the appropriate number of clusters and recognize their centers. Overall this research implies that using heterogeneous population in the genetic algorithm can lead to better results.
Eighteen scientists met at Jurata, Poland, to discuss various aspects of the transition from adolescence to adulthood. This transition is a delicate period facing complex interactions between the adolescents and the social group they belong to. Social identity, group identification and identity signalling, but also stress affecting basal salivary cortisol rhythms, hypertension, inappropriate nutrition causing latent and manifest obesity, moreover, in developing and under-developed countries, parasitosis causing anaemia thereby impairing growth and development, are issues to be dealt with during this period of the human development. In addition, some new aspects of the association between weight, height and head circumference in the newborns were discussed, as well as intrauterine head growth and head circumference as health risk indicators.
E-commerce marketplaces are highly dynamic with constant competition. While this competition is challenging for many merchants, it also provides plenty of opportunities, e.g., by allowing them to automatically adjust prices in order to react to changing market situations. For practitioners however, testing automated pricing strategies is time-consuming and potentially hazardously when done in production. Researchers, on the other side, struggle to study how pricing strategies interact under heavy competition. As a consequence, we built an open continuous time framework to simulate dynamic pricing competition called Price Wars. The microservice-based architecture provides a scalable platform for large competitions with dozens of merchants and a large random stream of consumers. Our platform stores each event in a distributed log. This allows to provide different performance measures enabling users to compare profit and revenue of various repricing strategies in real-time. For researchers, price trajectories are shown which ease evaluating mutual price reactions of competing strategies. Furthermore, merchants can access historical marketplace data and apply machine learning. By providing a set of customizable, artificial merchants, users can easily simulate both simple rule-based strategies as well as sophisticated data-driven strategies using demand learning to optimize their pricing strategies.
Cost models play an important role for the efficient implementation of software systems. These models can be embedded in operating systems and execution environments to optimize execution at run time. Even though non-uniform memory access (NUMA) architectures are dominating today's server landscape, there is still a lack of parallel cost models that represent NUMA system sufficiently. Therefore, the existing NUMA models are analyzed, and a two-step performance assessment strategy is proposed that incorporates low-level hardware counters as performance indicators. To support the two-step strategy, multiple tools are developed, all accumulating and enriching specific hardware event counter information, to explore, measure, and visualize these low-overhead performance indicators. The tools are showcased and discussed alongside specific experiments in the realm of performance assessment.
It has been observationally established that winds of hot massive stars have highly variable characteristics. The variability evident in the winds is believed to be caused by structures on a broad range of spatial scales. Small-scale structures (clumping) in stellar winds of hot stars are possible consequence of an instability appearing in their radiation hydrodynamics. To understand how clumping may influence calculation of theoretical spectra, different clumping properties and their 3D nature have to be taken into account. Properties of clumping have been examined using our 3D radiative transfer calculations. Effects of clumping for the case of the B[e] phenomenon are discussed.
Kijko et al. (2016) present various methods to estimate parameters that are relevant for probabilistic seismic-hazard assessment. One of these parameters, although not the most influential, is the maximum possible earthquake magnitude m(max). I show that the proposed estimation of m(max) is based on an erroneous equation related to a misuse of the estimator in Cooke (1979) and leads to unstable results. So far, reported finite estimations of m(max) arise from data selection, because the estimator in Kijko et al. (2016) diverges with finite probability. This finding is independent of the assumed distribution of earthquake magnitudes. For the specific choice of the doubly truncated Gutenberg-Richter distribution, I illustrate the problems by deriving explicit equations. Finally, I conclude that point estimators are generally not a suitable approach to constrain m(max).
The maximum entropy method is used to predict flows on water distribution networks. This analysis extends the water distribution network formulation of Waldrip et al. (2016) Journal of Hydraulic Engineering (ASCE), by the use of a continuous relative entropy defined on a reduced parameter set. This reduction in the parameters that the entropy is defined over ensures consistency between different representations of the same network. The performance of the proposed reduced parameter method is demonstrated with a one-loop network case study.
The Internet can be considered as the most important infrastructure for modern society and businesses. A loss of Internet connectivity has strong negative financial impacts for businesses and economies. Therefore, assessing Internet connectivity, in particular beyond their own premises and area of direct control, is of growing importance in the face of potential failures, accidents, and malicious attacks. This paper presents CORIA, a software framework for an easy analysis of connectivity risks based on large network graphs. It provides researchers, risk analysts, network managers and security consultants with a tool to assess an organization's connectivity and paths options through the Internet backbone, including a user-friendly and insightful visual representation of results. CORIA is flexibly extensible in terms of novel data sets, graph metrics, and risk scores that enable further use cases. The performance of CORIA is evaluated by several experiments on the Internet graph and further randomly generated networks.
Lately, first implementation approaches of Internet of Things (IoT) technologies penetrate industrial value-adding processes. Within this, the competence requirements for employees are changing. Employees’ organization, process, and interaction competences are of crucial importance in this new IoT environment, however, in students and vocational training not sufficiently considered yet. On the other hand, conventional learning factories evolve and transform to digital learning factories. Nevertheless, the integration of IoT technology and its usage for training in digital learning factories has been largely neglected thus far. Existing learning factories do not explicitly and properly consider IoT technology, which leads to deficiencies regarding an appropriate development of employees’ Industrial IoT competences. The goal of this contribution is to point out a didactic concept that enables development and training of these new demanded competences by using an IoT laboratory. For this purpose, a design science approach is applied. The result of this contribution is a didactic concept for the development of Industrial IoT competences in an IoT laboratory.
Dielectrophoretic functionalization of nanoelectrode arrays for the detection of influenza viruses
(2017)
Direct visualization of APLP1 cell-cell adhesion platforms via fluorescence fluctuation spectroscopy
(2017)
Dissolved CO2 storage in geological formations with low pressure, low risk and large capacities
(2017)
Geological CO2 storage is a mitigation technology to reduce CO2 emissions from fossil fuel combustion. However, major concerns are the pressure increase and saltwater displacement in the mainly targeted deep groundwater aquifers due to injection of supercritical CO2. The suggested solution is storage of CO2 exclusively in the dissolved state. In our exemplary regional case study of the North East German Basin based on a highly resolved temperature and pressure distribution model and a newly developed reactive transport coupling, we have quantified that 4.7 Gt of CO2 can be stored in solution compared to 1.5 Gt in the supercritical state.
This special issue is the result of several fruitful conference sessions on disturbance hydrology, which started at the 2013 AGU Fall Meeting in San Francisco and have continued every year since. The stimulating presentations and discussions surrounding those sessions have focused on understanding both the disruption of hydrologic functioning following discrete disturbances, as well as the subsequent recovery or change within the affected watershed system. Whereas some hydrologic disturbances are directly linked to anthropogenic activities, such as resource extraction, the contributions to this special issue focus primarily on those with indirect or less pronounced human involvement, such as bark-beetle infestation, wildfire, and other natural hazards. However, human activities are enhancing the severity and frequency of these seemingly natural disturbances, thereby contributing to acute hydrologic problems and hazards. Major research challenges for our increasingly disturbed planet include the lack of continuous pre and postdisturbance monitoring, hydrologic impacts that vary spatially and temporally based on environmental and hydroclimatic conditions, and the preponderance of overlapping or compounding disturbance sequences. In addition, a conceptual framework for characterizing commonalities and differences among hydrologic disturbances is still in its infancy. In this introduction to the special issue, we advance the fusion of concepts and terminology from ecology and hydrology to begin filling this gap. We briefly explore some preliminary approaches for comparing different disturbances and their hydrologic impacts, which provides a starting point for further dialogue and research progress.
DOES AGE INFLUENCE BRAIN POTENTIALS DURING AFFECTIVE PICTURE PROCESSING IN MIDDLE-AGED WOMEN?
(2017)
DPP4 inhibition prevents AKI
(2017)
Editorial
(2017)
Editorial
(2017)
Editorial
(2017)
Embedded smart home
(2017)
The popularity of MOOCs has increased considerably in the last years. A typical MOOC course consists of video content, self tests after a video and homework, which is normally in multiple choice format. After solving this homeworks for every week of a MOOC, the final exam certificate can be issued when the student has reached a sufficient score. There are also some attempts to include practical tasks, such as programming, in MOOCs for grading. Nevertheless, until now there is no known possibility to teach embedded system programming in a MOOC course where the programming can be done in a remote lab and where grading of the tasks is additionally possible. This embedded programming includes communication over GPIO pins to control LEDs and measure sensor values. We started a MOOC course called "Embedded Smart Home" as a pilot to prove the concept to teach real hardware programming in a MOOC environment under real life MOOC conditions with over 6000 students. Furthermore, also students with real hardware have the possibility to program on their own real hardware and grade their results in the MOOC course. Finally, we evaluate our approach and analyze the student acceptance of this approach to offer a course on embedded programming. We also analyze the hardware usage and working time of students solving tasks to find out if real hardware programming is an advantage and motivating achievement to support students learning success.
Emergency Care in Germany being re-assessed Hybrid Medical Care Model Seen As Potential Answer
(2017)
The design of embedded systems is becoming continuously more complex such that efficient system-level design methods are becoming crucial. Recently, combined Answer Set Programming (ASP) and Quantifier Free Integer Difference Logic (QF-IDL) solving has been shown to be a promising approach in system synthesis. However, this approach still has several restrictions limiting its applicability. In the paper at hand, we propose a novel ASP modulo Theories (ASPmT) system synthesis approach, which (i) supports more sophisticated system models, (ii) tightly integrates the QF-IDL solving into the ASP solving, and (iii) makes use of partial assignment checking. As a result, more realistic systems are considered and an early exclusion of infeasible solutions improves the entire system synthesis.
S-test results for the USGS and RELM forecasts. The differences between the simulated log-likelihoods and the observed log-likelihood are labelled on the horizontal axes, with scaling adjustments for the 40year.retro experiment. The horizontal lines represent the confidence intervals, within the 0.05 significance level, for each forecast and experiment. If this range contains a log-likelihood difference of zero, the forecasted log-likelihoods are consistent with the observed, and the forecast passes the S-test (denoted by thin lines). If the minimum difference within this range does not contain zero, the forecast fails the S-test for that particular experiment, denoted by thick lines. Colours distinguish between experiments (see Table 2 for explanation of experiment durations). Due to anomalously large likelihood differences, S-test results for Wiemer-Schorlemmer.ALM during the 10year.retro and 40year.retro experiments are not displayed. The range of log-likelihoods for the Holliday-et-al.PI forecast is lower than for the other forecasts due to relatively homogeneous forecasted seismicity rates and use of a small fraction of the RELM testing region.
We compare Visual Berrypicking, an interactive approach allowing users to explore large and highly faceted information spaces using similarity-based two-dimensional maps, with traditional browsing techniques. For large datasets, current projection methods used to generate maplike overviews suffer from increased computational costs and a loss of accuracy resulting in inconsistent visualizations. We propose to interactively align inexpensive small maps, showing local neighborhoods only, which ideally creates the impression of panning a large map. For evaluation, we designed a web-based prototype for movie exploration and compared it to the web interface of The Movie Database (TMDb) in an online user study. Results suggest that users are able to effectively explore large movie collections by hopping from one neighborhood to the next. Additionally, due to the projection of movie similarities, interesting links between movies can be found more easily, and thus, compared to browsing serendipitous discoveries are more likely.
As virtualization drives the automation of networking, the validation of security properties becomes more and more challenging eventually ruling out manual inspections. While formal verification in Software Defined Networks is provided by comprehensive tools with high speed reverification capabilities like NetPlumber for instance, the presence of middlebox functionality like firewalls is not considered. Also, they lack the ability to handle dynamic protocol elements like IPv6 extension header chains. In this work, we provide suitable modeling abstractions to enable both - the inclusion of firewalls and dynamic protocol elements. We exemplarily model the Linux ip6tables/netfilter packet filter and also provide abstractions for an application layer gateway. Finally, we present a prototype of our formal verification system FaVe.