Refine
Has Fulltext
- yes (668) (remove)
Year of publication
- 2015 (668) (remove)
Document Type
- Article (255)
- Postprint (197)
- Doctoral Thesis (108)
- Monograph/Edited Volume (33)
- Part of Periodical (26)
- Preprint (12)
- Review (11)
- Conference Proceeding (7)
- Master's Thesis (7)
- Bachelor Thesis (5)
Keywords
- Patholinguistik (20)
- Sprachtherapie (20)
- geistige Behinderung (20)
- mental deficiency (20)
- patholinguistics (20)
- primary progessive aphasia (20)
- primär progessive Aphasie (20)
- speech therapy (20)
- Armut (13)
- Nachhaltigkeit (13)
Institute
- Institut für Physik und Astronomie (108)
- Mathematisch-Naturwissenschaftliche Fakultät (70)
- Institut für Informatik und Computational Science (52)
- Institut für Chemie (39)
- MenschenRechtsZentrum (37)
- Humanwissenschaftliche Fakultät (36)
- Institut für Romanistik (32)
- Department Linguistik (30)
- Institut für Geowissenschaften (29)
- Institut für Biochemie und Biologie (24)
We investigate the ergodic properties of a random walker performing (anomalous) diffusion on a random fractal geometry. Extensive Monte Carlo simulations of the motion of tracer particles on an ensemble of realisations of percolation clusters are performed for a wide range of percolation densities. Single trajectories of the tracer motion are analysed to quantify the time averaged mean squared displacement (MSD) and to compare this with the ensemble averaged MSD of the particle motion. Other complementary physical observables associated with ergodicity are studied, as well. It turns out that the time averaged MSD of individual realisations exhibits non-vanishing fluctuations even in the limit of very long observation times as the percolation density approaches the critical value. This apparent non-ergodic behaviour concurs with the ergodic behaviour on the ensemble averaged level. We demonstrate how the non-vanishing fluctuations in single particle trajectories are analytically expressed in terms of the fractal dimension and the cluster size distribution of the random geometry, thus being of
purely geometrical origin. Moreover, we reveal that the convergence scaling law to ergodicity, which is known to be inversely proportional to the observation time T for ergodic diffusion processes, follows a power-law BT� h with h o 1 due to the fractal structure of the accessible space. These results provide useful measures for differentiating the subdiffusion on random fractals from an otherwise closely related process, namely, fractional Brownian motion. Implications of our results on the analysis of single particle tracking experiments are provided.
The interruption of learning processes by breaks filled with diverse activities is common in everyday life. We investigated the effects of active computer gaming and passive relaxation (rest and music) breaks on working memory performance. Young adults were exposed to breaks involving (i) eyes-open resting, (ii) listening to music and (iii) playing the video game “Angry Birds” before performing the n-back working memory task. Based on linear mixed-effects modeling, we found that playing the “Angry Birds” video game during a short learning break led to a decline in task performance over the course of the task as compared to eyes-open resting and listening to music, although overall task performance was not impaired. This effect was associated with high levels of daily mind wandering and low self-reported ability to concentrate. These findings indicate that video games can negatively affect working memory performance over time when played in between learning tasks. We suggest further investigation of these effects because of their relevance to everyday activity.
The Brazilian Cerrado is recognised as one of the most threatened biomes in the world, as the region has experienced a striking change from natural vegetation to intense cash crop production. The impacts of rapid agricultural expansion on soil and water resources are still poorly understood in the region. Therefore, the overall aim of the thesis is to improve our understanding of the ecohydrological processes causing water and soil degradation in the Brazilian Cerrado.
I first present a metaanalysis to provide quantitative evidence and identifying the main impacts of soil and water alterations resulting from land use change. Second, field studies were conducted to (i) examine the effects of land use change on soils of natural cerrado transformed to common croplands and pasture and (ii) indicate how agricultural production affects water quality across a meso-scale catchment. Third, the ecohydrological process-based model SWAT was tested with simple scenario analyses to gain insight into the impacts of land use and climate change on the water cycling in the upper São Lourenço catchment which experienced decreasing discharges in the last 40 years.
Soil and water quality parameters from different land uses were extracted from 89 soil and 18 water studies in different regions across the Cerrado. Significant effects on pH, bulk density and available P and K for croplands and less-pronounced effects on pastures were evident. Soil total N did not differ between land uses because most of the cropland sites were N-fixing soybean cultivations, which are not artificially fertilized with N. By contrast, water quality studies showed N enrichment in agricultural catchments, indicating fertilizer impacts and potential susceptibility to eutrophication. Regardless of the land use, P is widely absent because of the high-fixing capacities of deeply weathered soils and the filtering capacity of riparian vegetation. Pesticides, however, were consistently detected throughout the entire aquatic system. In several case studies, extremely high-peak concentrations exceeded Brazilian and EU water quality limits, which pose serious health risks.
My field study revealed that land conversion caused a significant reduction in infiltration rates near the soil surface of pasture (–96 %) and croplands (–90 % to –93 %). Soil aggregate stability was significantly reduced in croplands than in cerrado and pasture. Soybean crops had extremely high extractable P (80 mg kg–1), whereas pasture N levels declined. A snapshot water sampling showed strong seasonality in water quality parameters. Higher temperature, oxi-reduction potential (ORP), NO2–, and very low oxygen concentrations (<5 mg•l–1) and saturation (<60 %) were recorded during the rainy season. By contrast, remarkably high PO43– concentrations (up to 0.8 mg•l–1) were measured during the dry season. Water quality parameters were affected by agricultural activities at all sampled sub-catchments across the catchment, regardless of stream characteristic. Direct NO3– leaching appeared to play a minor role; however, water quality is affected by topsoil fertiliser inputs with impact on small low order streams and larger rivers. Land conversion leaving cropland soils more susceptible to surface erosion by increased overland flow events.
In a third study, the field data were used to parameterise SWAT. The model was tested with different input data and calibrated in SWAT-CUP using the SUFI-2 algorithm. The model was judged reliable to simulate the water balance in the Cerrado. A complete cerrado, pasture and cropland cover was used to analyse the impact of land use on water cycling as well as climate change projections (2039–2058) according to the projections of the RCP 8.5 scenario. The actual evapotranspiration (ET) for the cropland scenario was higher compared to the cerrado cover (+100 mm a–1). Land use change scenarios confirmed that deforestation caused higher annual ET rates explaining partly the trend of decreased streamflow. Taking all climate change scenarios into account, the most likely effect is a prolongation of the dry season (by about one month), with higher peak flows in the rainy season. Consequently, potential threats for crop production with lower soil moisture and increased erosion and sediment transport during the rainy season are likely and should be considered in adaption plans.
From the three studies of the thesis I conclude that land use intensification is likely to seriously limit the Cerrado’s future regarding both agricultural productivity and ecosystem stability. Because only limited data are available for the vast biome, we recommend further field studies to understand the interaction between terrestrial and aquatic systems. This thesis may serve as a valuable database for integrated modelling to investigate the impact of land use and climate change on soil and water resources and to test and develop mitigation measures for the Cerrado in the future.
Nowadays, business processes are increasingly supported by IT services that produce massive amounts of event data during process execution. Aiming at a better process understanding and improvement, this event data can be used to analyze processes using process mining techniques. Process models can be automatically discovered and the execution can be checked for conformance to specified behavior. Moreover, existing process models can be enhanced and annotated with valuable information, for example for performance analysis. While the maturity of process mining algorithms is increasing and more tools are entering the market, process mining projects still face the problem of different levels of abstraction when comparing events with modeled business activities. Mapping the recorded events to activities of a given process model is essential for conformance checking, annotation and understanding of process discovery results. Current approaches try to abstract from events in an automated way that does not capture the required domain knowledge to fit business activities. Such techniques can be a good way to quickly reduce complexity in process discovery. Yet, they fail to enable techniques like conformance checking or model annotation, and potentially create misleading process discovery results by not using the known business terminology.
In this thesis, we develop approaches that abstract an event log to the same level that is needed by the business. Typically, this abstraction level is defined by a given process model. Thus, the goal of this thesis is to match events from an event log to activities in a given process model. To accomplish this goal, behavioral and linguistic aspects of process models and event logs as well as domain knowledge captured in existing process documentation are taken into account to build semiautomatic matching approaches. The approaches establish a pre--processing for every available process mining technique that produces or annotates a process model, thereby reducing the manual effort for process analysts. While each of the presented approaches can be used in isolation, we also introduce a general framework for the integration of different matching approaches.
The approaches have been evaluated in case studies with industry and using a large industry process model collection and simulated event logs. The evaluation demonstrates the effectiveness and efficiency of the approaches and their robustness towards nonconforming execution logs.
Business process management (BPM) is a systematic and structured approach to model, analyze, control, and execute business operations also referred to as business processes that get carried out to achieve business goals. Central to BPM are conceptual models. Most prominently, process models describe which tasks are to be executed by whom utilizing which information to reach a business goal. Process models generally cover the perspectives of control flow, resource, data flow, and information systems.
Execution of business processes leads to the work actually being carried out. Automating them increases the efficiency and is usually supported by process engines. This, though, requires the coverage of control flow, resource assignments, and process data. While the first two perspectives are well supported in current process engines, data handling needs to be implemented and maintained manually. However, model-driven data handling promises to ease implementation, reduces the error-proneness through graphical visualization, and reduces development efforts through code generation.
This thesis addresses the modeling, analysis, and execution of data in business processes and presents a novel approach to execute data-annotated process models entirely model-driven. As a first step and formal grounding for the process execution, a conceptual framework for the integration of processes and data is introduced. This framework is complemented by operational semantics through a Petri net mapping extended with data considerations. Model-driven data execution comprises the handling of complex data dependencies, process data, and data exchange in case of communication between multiple process participants. This thesis introduces concepts from the database domain into BPM to enable the distinction of data operations, to specify relations between data objects of the same as well as of different types, to correlate modeled data nodes as well as received messages to the correct run-time process instances, and to generate messages for inter-process communication. The underlying approach, which is not limited to a particular process description language, has been implemented as proof-of-concept.
Automation of data handling in business processes requires data-annotated and correct process models. Targeting the former, algorithms are introduced to extract information about data nodes, their states, and data dependencies from control information and to annotate the process model accordingly. Usually, not all required information can be extracted from control flow information, since some data manipulations are not specified. This requires further refinement of the process model. Given a set of object life cycles specifying allowed data manipulations, automated refinement of the process model towards containment of all data manipulations is enabled. Process models are an abstraction focusing on specific aspects in detail, e.g., the control flow and the data flow views are often represented through activity-centric and object-centric process models. This thesis introduces algorithms for roundtrip transformations enabling the stakeholder to add information to the process model in the view being most appropriate.
Targeting process model correctness, this thesis introduces the notion of weak conformance that checks for consistency between given object life cycles and the process model such that the process model may only utilize data manipulations specified directly or indirectly in an object life cycle. The notion is computed via soundness checking of a hybrid representation integrating control flow and data flow correctness checking. Making a process model executable, identified violations must be corrected. Therefore, an approach is proposed that identifies for each violation multiple, alternative changes to the process model or the object life cycles.
Utilizing the results of this thesis, business processes can be executed entirely model-driven from the data perspective in addition to the control flow and resource perspectives already supported before. Thereby, the model creation is supported by algorithms partly automating the creation process while model consistency is ensured by data correctness checks.
In diesem Papier wird das Konzept eines Lernzentrums für die Informatik (LZI) an der Universität Paderborn vorgestellt. Ausgehend von den fachspezifischen Schwierigkeiten der Informatik Studierenden werden die Angebote des LZIs erläutert, die sich über die vier Bereiche Individuelle Beratung und Betreuung, „Offener Lernraum“, Workshops und Lehrveranstaltungen sowie Forschung erstrecken. Eine erste Evaluation mittels Feedbackbögen zeigt, dass das Angebot bei den Studierenden positiv aufgenommen wird. Zukünftig soll das Angebot des LZIs weiter ausgebaut und verbessert werden. Ausgangsbasis dazu sind weitere Studien.
Die Wahl des richtigen Studienfaches und die daran anschließende
Studieneingangsphase sind oft entscheidend für den erfolgreichen Verlauf eines Studiums. Eine große Herausforderung besteht dabei darin, bereits in den ersten Wochen des Studiums bestehende Defizite in vermeintlich einfachen Schlüsselkompetenzen zu erkennen und diese so bald wie möglich zu beheben. Eine zweite, nicht minder wichtige Herausforderung ist es, möglichst frühzeitig für jeden einzelnen Studierenden zu erkennen, ob er bzw. sie das individuell richtige Studienfach gewählt hat, das den jeweiligen persönlichen Neigungen, Interessen und Fähigkeiten entspricht und zur Verwirklichung der eigenen Lebensziele beiträgt. Denn nur dann sind Studierende ausreichend stark und dauerhaft intrinsisch motiviert, um ein anspruchsvolles, komplexes Studium erfolgreich durchzuziehen. In diesem Beitrag fokussieren wir eine Maßnahme, die die Studierenden an einen Prozess zur systematischen Reflexion des eigenen Lernprozesses und der eigenen Ziele heranführt und beides in Relation setzt.
Ziel einer neuen Studieneingangsphase ist, den Studierenden bis zum Ende des ersten Semesters ein vielfältiges Berufsbild der Informatik und Wirtschaftsinformatik mit dem breiten Aufgabenspektrum aufzublättern und damit die Zusammenhänge zwischen den einzelnen Modulen des Curriculums zu verdeutlichen. Die Studierenden sollen in die Lage versetzt werden, sehr eigenständig die Planung und Gestaltung ihres Studiums in die Hand zu nehmen.
Es wird ein Informatik-Wettbewerb für Schülerinnen und Schüler der Sekundarstufe II beschrieben, der über mehrere Wochen möglichst realitätsnah die Arbeitswelt eines Informatikers vorstellt. Im Wettbewerb erarbeiten die Schülerteams eine Android-App und organisieren ihre Entwicklung durch Projektmanagementmethoden, die sich an professionellen, agilen Prozessen orientieren. Im Beitrag werden der theoretische Hintergrund zu Wettbewerben, die organisatorischen und didaktischen Entscheidung, eine erste Evaluation sowie Reflexion und Ausblick dargestellt.
In der Lehre zur MCI (Mensch-Computer-Interaktion) stellt sich immer wieder die Herausforderung, praktische Übungen mit spannenden Ergebnissen durchzuführen, die sich dennoch nicht in technischen Details verlieren sondern MCI-fokussiert bleiben. Im Lehrmodul „Interaktionsdesign“ an der Universität Hamburg werden von Studierenden innerhalb von drei Wochen prototypische Interaktionskonzepte für das Spiel Neverball entworfen und praktisch umgesetzt. Anders als in den meisten Grundlagenkursen zur MCI werden hier nicht Mock-Ups, sondern lauffähige Software entwickelt. Um dies innerhalb der Projektzeit zu ermöglichen, wurde Neverball um eine TCP-basierte Schnittstelle erweitert. So entfällt die aufwändige Einarbeitung in den Quellcode des Spiels und die Studierenden können sich auf ihre Interaktionsprototypen konzentrieren. Wir beschreiben die Erfahrungen aus der
mehrmaligen Durchführung des Projektes und erläutern unser Vorgehen bei der Umsetzung. Die Ergebnisse sollen Lehrende im Bereich MCI unterstützen, ähnliche praxisorientierte Übungen mit Ergebnissen „zum Anfassen“ zu gestalten.
Der folgende Artikel beschreibt die Evaluation eines Lehrvideos zum informatischen Problemlösen, welches auf der Grundlage einer Vergleichsstudie mit starken und schwachen Problemlösern entwickelt wurde. Beispielhaft wird in dem Film ein Färbeproblem durch einen fiktiven Hochleister unter lautem Denken gelöst, die einzelnen Arbeitsschritte werden abschnittsweise kommentiert und erklärt. Ob dieses Lernkonzept von Studenten akzeptiert wird und sich durch Anschauen des Videos tatsächlich ein Lerneffekt einstellt, wurde durch eine Befragung und eine erste Vergleichsstudie untersucht.
Über die Autoren
(2015)
Auf der Grundlage der Planung, Durchführung, Evaluation und Revision eines gemeinsamen Seminars von Medienpädagogik und Didaktik der Informatik stellen wir in diesem Aufsatz dar, wo die Defizite klassischer Medienbildung in Bezug auf digitale bzw. interaktive Medien liegen und welche Inhalte der Informatik für Studierende aller Lehrämter – im allgemeinbildenden Sinne – aus dieser Perspektive relevant erscheinen.
When it comes to footnotes, Alexander von Humboldt was ahead of his times even though his references leave much to be desired by today’s academic standards. This article examines the footnotes of Humboldt’s Essai politique sur l‘île de Cuba (1826). While it is not always easy to decipher his sometimes cryptic references, the undertaking is worthwhile: Humboldt’s footnotes do not only reveal his vast networks of knowledge. They also provide glimpses of ongoing, contemporary disputes among different scholars that involve Humboldt’s writings. They also present Humboldt’s reactions to such disputes. Exploring Humboldt’s footnotes consequently allows the reader to access both Humboldt the scholar and Humboldt the human being.
Gustav Rose (1798-1873) begleitete Alexander von Humboldt auf seiner Russlandreise und stand bis zu Humboldts Tod persönlich und postalisch zum ihm in Kontakt. Die Edition des vorliegenden Briefs zielt darauf ab, die Bedeutung der Person Gustav Rose in ihrer Beziehung zu Alexander von Humboldt und ihrem Einfluss auf den mineralogisch-geologischen Teil des Kosmos zu beleuchten und dem Leser dieses interessante historische Dokument zugänglich zu machen.
Continental rifts are excellent regions where the interplay between extension, the build-up of topography, erosion and sedimentation can be evaluated in the context of landscape evolution. Rift basins also constitute important archives that potentially record the evolution and migration of species and the change of sedimentary conditions as a result of climatic change. Finally, rifts have increasingly become targets of resource exploration, such as hydrocarbons or geothermal systems. The study of extensional processes and the factors that further modify the mainly climate-driven surface process regime helps to identify changes in past and present tectonic and geomorphic processes that are ultimately recorded in rift landscapes.
The Cenozoic East African Rift System (EARS) is an exemplary continental rift system and ideal natural laboratory to observe such interactions. The eastern and western branches of the EARS constitute first-order tectonic and topographic features in East Africa, which exert a profound influence on the evolution of topography, the distribution and amount of rainfall, and thus the efficiency of surface processes. The Kenya Rift is an integral part of the eastern branch of the EARS and is characterized by high-relief rift escarpments bounded by normal faults, gently tilted rift shoulders, and volcanic centers along the rift axis.
Considering the Cenozoic tectonic processes in the Kenya Rift, the tectonically controlled cooling history of rift shoulders, the subsidence history of rift basins, and the sedimentation along and across the rift, may help to elucidate the morphotectonic evolution of this extensional province. While tectonic forcing of surface processes may play a minor role in the low-strain rift on centennial to millennial timescales, it may be hypothesized that erosion and sedimentation processes impacted by climate shifts associated with pronounced changes in the availability in moisture may have left important imprints in the landscape.
In this thesis I combined thermochronological, geomorphic field observations, and morphometry of digital elevation models to reconstruct exhumation processes and erosion rates, as well as the effects of climate on the erosion processes in different sectors of the rift. I present three sets of results: (1) new thermochronological data from the northern and central parts of the rift to quantitatively constrain the Tertiary exhumation and thermal evolution of the Kenya Rift. (2) 10Be-derived catchment-wide mean denudation rates from the northern, central and southern rift that characterize erosional processes on millennial to present-day timescales; and (3) paleo-denudation rates in the northern rift to constrain climatically controlled shifts in paleoenvironmental conditions during the early Holocene (African Humid Period).
Taken together, my studies show that time-temperature histories derived from apatite fission track (AFT) analysis, zircon (U-Th)/He dating, and thermal modeling bracket the onset of rifting in the Kenya Rift between 65-50 Ma and about 15 Ma to the present. These two episodes are marked by rapid exhumation and, uplift of the rift shoulders. Between 45 and 15 Ma the margins of the rift experienced very slow erosion/exhumation, with the accommodation of sediments in the rift basin.
In addition, I determined that present-day denudation rates in sparsely vegetated parts of the Kenya Rift amount to 0.13 mm/yr, whereas denudation rates in humid and more densely vegetated sectors of the rift flanks reach a maximum of 0.08 mm/yr, despite steeper hillslopes. I inferred that hillslope gradient and vegetation cover control most of the variation in denudation rates across the Kenya Rift today. Importantly, my results support the notion that vegetation cover plays a fundamental role in determining the voracity of erosion of hillslopes through its stabilizing effects on the land surface.
Finally, in a pilot study I highlighted how paleo-denudation rates in climatic threshold areas changed significantly during times of transient hydrologic conditions and involved a sixfold increase in erosion rates during increased humidity. This assessment is based on cosmogenic nuclide (10Be) dating of quartzitic deltaic sands that were deposited in the northern Kenya Rift during a highstand of Lake Suguta, which was associated with the Holocene African Humid Period. Taken together, my new results document the role of climate variability in erosion processes that impact climatic threshold environments, which may provide a template for potential future impacts of climate-driven changes in surface processes in the course of Global Change.
Spots on stellar surfaces are thought to be stellar analogues of sunspots. Thus, starspots are direct manifestations of strong magnetic fields. Their decay rate is directly related to the magnetic diffusivity, which itself is a key quantity for the deduction of an activity cycle length. So far, no single starspot decay has been observed, and thus no stellar activity cycle was inferred from its corresponding turbulent diffusivity.
We investigate the evolution of starspots on the rapidly-rotating K0 giant XX Triangulum. Continuous high-resolution and phase-resolved spectroscopy was obtained with the robotic 1.2-m STELLA telescope on Tenerife over a timespan of six years. With our line-profile inversion code iMap we reconstruct a total of 36 consecutive Doppler maps. To quantify starspot area decay and growth, we match the observed images with simplified spot models based on a Monte-Carlo approach.
It is shown that the surface of XX Tri is covered with large high-latitude and even polar spots and with occasional small equatorial spots. Just over the course of six years, we see a systematically changing spot distribution with various time scales and morphology such as spot fragmentation and spot merging as well as spot decay and formation.
For the first time, a starspot decay rate on another star than the Sun is determined. From our spot-decay analysis we determine an average linear decay rate of D = -0.067±0.006 Gm^2/day. From this decay rate, we infer a turbulent diffusivity of η_τ = (6.3±0.5) x 10^14 cm^2/s and consequently predict an activity cycle of 26±6 years. The obtained cycle length matches very well with photometric observations.
Our time-series of Doppler maps further enables to investigate the differential rotation of XX Tri. We therefore applied a cross-correlation analysis. We detect a weak solar-like differential rotation with a surface shear of α = 0.016±0.003. This value agrees with similar studies of other RS CVn stars.
Furthermore, we found evidence for active longitudes and flip-flops. Whereas the more active longitude is located in phase towards the (unseen) companion star, the weaker active longitude is located at the opposite stellar hemisphere. From their periodic appearance, we infer a flip-flop cycle of ~2 years. Both activity phenomena are common on late-type binary stars.
Last but not least we redetermine several astrophysical properties of XX Tri and its binary system, as large datasets of photometric and spectroscopic observations are available since its last determination in 1999. Additionally, we compare the rotational spot-modulation from photometric and spectroscopic studies.
The Barberton Greenstone Belt (BGB) in the northwestern part of South Africa belongs to the few well-preserved remnants of Archean crust. Over the last centuries, the BGB has been intensively studied at surface with detailed mapping of its surfacial geological units and tectonic features. Nevertheless, the deeper structure of the BGB remains poorly understood. Various tectonic evolution models have been developed based on geo-chronological and structural data. These theories are highly controversial and centre on the question whether plate tectonics - as geoscientists understand them today - was already evolving on the Early Earth or whether vertical mass movements driven by the higher temperature of the Earth in Archean times governed continent development.
To get a step closer to answering the questions regarding the internal structure and formation of the BGB, magnetotelluric (MT) field experiments were conducted as part of the German-South African research initiative Inkaba yeAfrica. Five-component MT data (three magnetic and two electric channels) were collected at ~200 sites aligned along six profiles crossing the southern part of the BGB. Tectonic features like (fossil) faults and shear zones are often mineralized and therefore can have high electrical conductivities. Hence, by obtaining an image of the conductivity distribution of the subsurface from MT measurements can provide useful information on tectonic processes.
Unfortunately, the BGB MT data set is heavily affected by man-made electromagnetic noise caused, e.g. by powerlines and electric fences. Aperiodic spikes in the magnetic and corresponding offsets in the electric field components impair the data quality particularly at periods >1 s which are required to image deep electrical structures. Application of common methods for noise reduction like delay filtering and remote reference processing, only worked well for periods <1 s. Within the framework of this thesis two new filtering approaches were developed to handle the severe noise in long period data and obtain reliable processing results. The first algorithm is based on the Wiener filter in combination with a spike detection algorithm. Comparison of data variances of a local site with those of a reference site allows the identification of disturbed time series windows for each recorded channel at the local site. Using the data of the reference site, a Wiener filter algorithm is applied to predict physically meaningful data to replace the disturbed windows. While spikes in the magnetic channels are easily recognized and replaced, steps in the electric channels are more difficult to detect depending on their offset. Therefore, I have implemented a novel approach based on time series differentiation, noise removal and subsequent integration to overcome this obstacle. A second filtering approach where spikes and steps in the time series are identified using a comparison of the short and long time average of the data was also implemented as part of my thesis. For this filtering approach the noise in the form of spikes and offsets in the data is treated by an interpolation of the affected data samples. The new developments resulted in a substantial data improvement and allowed to gain one to two decades of data (up to 10 or 100 s).
The re-processed MT data were used to image the electrical conductivity distribution of the BGB by 2D and 3D inversion. Inversion models are in good agreement with the surface geology delineating the highly resistive rocks of the BGB from surrounding more conductive geological units. Fault zones appear as conductive structures and can be traced to depths of 5 to 10 km. 2D models suggest a continuation of the faults further south across the boundary of the BGB. Based on the shallow tectonic structures (fault system) within the BGB compared to deeply rooted resistive batholiths in the area, tectonic models including both vertical mass transport and in parts present-day style plate tectonics seem to be most likely for the evolution of the BGB.
Alexander von Humboldts 1827/28 gehaltene Kosmos-Vorträge stellen ebenso einen kanonischen Bezugspunkt wie einen weißen Fleck in der Rezeption seines Schaffens dar. Zahlreiche Deutungen basieren auf lediglich zwei unkommentierten Leseausgaben von Collegheften anonym gebliebener Zuhörer Humboldts. Seine 2009 wiedergefundenen eigenhändigen Vorlesungsmanuskripte versprechen, diese schmale Quellenbasis zu erweitern. Alfred Dove hatte bereits 1872 auf ihre Existenz verwiesen und interpretierte sie damals als Nukleus der schriftlichen Ausarbeitung des Kosmos. Diese Sichtweise hat sich in der Rezeption gehalten, während die Materialien, die Dove beschrieb, in Vergessenheit gerieten. Der Artikel stellt anhand ausgewählter Beispiele die Vorlesungsmanuskripte Humboldts und Nachschriften seiner Zuhörer in ihrem Zusammenhang vor, entwickelt die These von der Eigenständigkeit der sogenannten Kosmos-Vorträge gegenüber dem fünfbändigen Kosmos und umreißt die wichtigsten Ziele und Inhalte der geplanten digitalen Edition.
In this paper we discuss how Alexander von Humboldt conceived a past to New Spain in his Political Essay on New Spain (1811) and how this text was, in turn, appropriated by the Mexican historiography during the 19th century.
In order to do so, we analyze how the Prussian drew from American sources, particularly from the text of the Jesuit Francisco Javier Clavijero, written shortly before. We also study Humboldt’s conceptions of text and of history, highlighting the place of the indigenous in the composition of his reasoning. Finally, we give examples of how the Mexican nationalist historiography read and reinterpreted the Political Essay.
Este artículo explora la recepción realizada por Alexander von Humboldt de la figura de Cristóbal Colón, principalmente en su Relation historique y en su Examen critique. En primer lugar, el artículo explora lo que los biógrafos de Humboldt han llamado “ein zweiter Kolumbus” (un segundo Colón). En segundo lugar, se traza la historia de la edificación humboldtianna de una “imaginación poética” y de una “geografía mítica” atribuidas por Humboldt a Cristóbal Colón.
Spectral fingerprinting
(2015)
Current research on runoff and erosion processes, as well as an increasing demand for sustainable watershed management emphasize the need for an improved understanding of sediment dynamics. This involves the accurate assessment of erosion rates and sediment transfer, yield and origin. A variety of methods exist to capture these processes at the catchment scale. Among these, sediment fingerprinting, a technique to trace back the origin of sediment, has attracted increasing attention by the scientific community in recent years. It is a two-step procedure, based on the fundamental assumptions that potential sources of sediment can be reliably discriminated based on a set of characteristic ‘fingerprint’ properties, and that a comparison of source and sediment fingerprints allows to quantify the relative contribution of each source.
This thesis aims at further assessing the potential of spectroscopy to assist and improve the sediment fingerprinting technique. Specifically, this work focuses on (1) whether potential sediment sources can be reliably identified based on spectral features (‘fingerprints’), whether (2) these spectral fingerprints permit the quantification of relative source contribution, and whether (3) in situ derived source information is sufficient for this purpose. Furthermore, sediment fingerprinting using spectral information is applied in a study catchment to (4) identify major sources and observe how relative source contributions change between and within individual flood events. And finally, (5) spectral fingerprinting results are compared and combined with simultaneous sediment flux measurements to study sediment origin, transport and storage behaviour.
For the sediment fingerprinting approach, soil samples were collected from potential sediment sources within the Isábena catchment, a meso-scale basin in the central Spanish Pyrenees. Undisturbed samples of the upper soil layer were measured in situ using an ASD spectroradiometer and subsequently sampled for measurements in the laboratory. Suspended sediment was sampled automatically by means of ISCO samplers at the catchment as well as at the five major subcatchment outlets during flood events, and stored fine sediment from the channel bed was collected from 14 cross-sections along the main river. Artificial mixtures of known contributions were produced from source soil samples. Then, all source, sediment and mixture samples were dried and spectrally measured in the laboratory. Subsequently, colour coefficients and physically based features with relation to organic carbon, iron oxide, clay content and carbonate, were calculated from all in situ and laboratory spectra. Spectral parameters passing a number of prerequisite tests were submitted to principal component analyses to study natural clustering of samples, discriminant function analyses to observe source differentiation accuracy, and a mixing model for source contribution assessment. In addition, annual as well as flood event based suspended sediment fluxes from the catchment and its subcatchments were calculated from rainfall, water discharge and suspended sediment concentration measurements using rating curves and Quantile Regression Forests. Results of sediment flux monitoring were interpreted individually with respect to storage behaviour, compared to fingerprinting source ascriptions and combined with fingerprinting to assess their joint explanatory potential.
In response to the key questions of this work, (1) three source types (land use) and five spatial sources (subcatchments) could be reliably discriminated based on spectral fingerprints. The artificial mixture experiment revealed that while (2) laboratory parameters permitted source contribution assessment, (3) the use of in situ derived information was insufficient. Apparently, high discrimination accuracy does not necessarily imply good quantification results. When applied to suspended sediment samples of the catchment outlet, the spectral fingerprinting approach was able to (4) quantify the major sediment sources: badlands and the Villacarli subcatchment, respectively, were identified as main contributors, which is consistent with field observations and previous studies. Thereby, source contribution was found to vary both, within and between individual flood events. Also sediment flux was found to vary considerably, annually as well as seasonally and on flood event base. Storage was confirmed to play an important role in the sediment dynamics of the studied catchment, whereas floods with lower total sediment yield tend to deposit and floods with higher yield rather remove material from the channel bed. Finally, a comparison of flux measurements with fingerprinting results highlighted the fact that (5) immediate transport from sources to the catchment outlet cannot be assumed. A combination of the two methods revealed different aspects of sediment dynamics that none of the techniques could have uncovered individually.
In summary, spectral properties provide a fast, non-destructive, and cost-efficient means to discriminate and quantify sediment sources, whereas, unfortunately, straight-forward in situ collected source information is insufficient for the approach. Mixture modelling using artificial mixtures permits valuable insights into the capabilities and limitations of the method and similar experiments are strongly recommended to be performed in the future. Furthermore, a combination of techniques such as e.g. (spectral) sediment fingerprinting and sediment flux monitoring can provide comprehensive understanding of sediment dynamics.
The Norway lobster, Nephrops norvegicus, is a burrowing decapod with a rhythmic burrow emergence (24 h) governed by the circadian system. It is an important resource for European fisheries and its behavior deeply affects its availability. The current knowledge of Nephrops circadian biology is phenomenological as it is currently the case for almost all crustaceans. In attempt to elucidate the putative molecular mechanisms underlying circadian gene regulation in Nephrops, we used a transcriptomics approach on cDNA extracted from the eyestalk, a structure playing a crucial role in controlling behavior of decapods. We studied 14 male lobsters under 12–12 light-darkness blue light cycle. We used the Hiseq 2000 Illumina platform to sequence two eyestalk libraries (under light and darkness conditions) obtaining about 90 millions 100-bp paired-end reads. Trinity was used for the de novo reconstruction of transcriptomes; the size at which half of all assembled bases reside in contigs (N50) was equal to 1796 (light) and 2055 (darkness). We found a list of candidate clock genes and focused our attention on canonical ones: timeless, period, clock and bmal1. The cloning of assembled fragments validated Trinity outputs. The putative Nephrops clock genes showed high levels of identity (blastx on NCBI) with known crustacean clock gene homologs such as Eurydice pulchra (period: 47%, timeless: 59%, bmal1: 79%) and Macrobrachium rosenbergii (clock: 100%). We also found a vertebrate-like cryptochrome 2. RT-qPCR showed that only timeless had a robust diel pattern of expression. Our data are in accordance with the current knowledge of the crustacean circadian clock, reinforcing the idea that the molecular clockwork of this group shows some differences with the established model in Drosophila melanogaster.
Climate change is likely to impact the seasonality and generation processes of floods in the Nordic countries, which has direct implications for flood risk assessment, design flood estimation, and hydropower production management. Using a multi-model/multi-parameter approach to simulate daily discharge for a reference (1961–1990) and a future (2071–2099) period, we analysed the projected changes in flood seasonality and generation processes in six catchments with mixed snowmelt/rainfall regimes under the current climate in Norway. The multi-model/multi-parameter ensemble consists of (i) eight combinations of global and regional climate models, (ii) two methods for adjusting the climate model output to the catchment scale, and (iii) one conceptual hydrological model with 25 calibrated parameter sets. Results indicate that autumn/winter events become more frequent in all catchments considered, which leads to an intensification of the current autumn/winter flood regime for the coastal catchments, a reduction of the dominance of spring/summer flood regimes in a high-mountain catchment, and a possible systematic shift in the current flood regimes from spring/summer to autumn/winter in the two catchments located in northern and south-eastern Norway. The changes in flood regimes result from increasing event magnitudes or frequencies, or a combination of both during autumn and winter. Changes towards more dominant autumn/winter events correspond to an increasing relevance of rainfall as a flood generating process (FGP) which is most pronounced in those catchments with the largest shifts in flood seasonality. Here, rainfall replaces snowmelt as the dominant FGP primarily due to increasing temperature.We further analysed the ensemble components in contributing to overall uncertainty in the projected changes and found that the climate projections and the methods for downscaling or bias correction tend to be the largest contributors. The relative role of hydrological parameter uncertainty, however, is highest for those catchments showing the largest changes in flood seasonality, which confirms the lack of robustness in hydrological model parameterization for simulations under transient hydrometeorological conditions.
Der kurze Aufsatz gibt einen Überblick über den erhaltenen Briefwechsel zwischen A. v. Humboldt und Charles Lyell. Er veröffentlicht darüber hinaus einen undatierten, bisher unbekannten Brief Humboldts an Lyell. Der Brief kann mit Hilfe eines Briefes an Pertz datiert werden. Deshalb wird auch dieser Brief zum ersten Mal vollständig veröffentlicht.
Zwischen Sein und Sollen
(2015)
Bericht über die Tätigkeit des Menschenrechtsausschusses der Vereinten Nationen im Jahre 2014
(2015)
Zusammenfassung – nichtamtliche Leitsätze:
• Ein allgemeines Burkaverbot in der Öffentlichkeit verstößt nicht gegen Art. 8 oder 9
EMRK.
• Das Tragen einer Burka unterfällt als private Lebensführung dem Schutzbereich des
Art. 8 EMRK sowie als Religionsausübung dem Schutzbereich des Art. 9 EMRK.
• Ein allgemeines Burkaverbot in der Öffentlichkeit zum Schutz der Sicherheit verfolgt
ein legitimes Ziel, ist in Ermangelung einer allgemeinen Gefahr aber nicht erforderlich
und damit unverhältnismäßig.
• Der Schutz des Respekts für die Gleichheit von Mann und Frau oder für die Menschenwürde
lässt sich nicht als legitimes Ziel für ein Burkaverbot heranziehen.
• Der Schutz des Respekts für die Minimalbedingungen gesellschaftlichen Zusammenlebens
stellt ein legitimes Ziel für ein Burkaverbot dar. Das Burkaverbot ist angesichts
des breiten staatlichen Ermessensspielraums in einer solchen Frage nicht unverhältnismäßig.
Microsaccades
(2015)
The first thing we do upon waking is open our eyes. Rotating them in our eye sockets, we scan our surroundings and collect the information into a picture in our head. Eye movements can be split into saccades and fixational eye movements, which occur when we attempt to fixate our gaze. The latter consists of microsaccades, drift and tremor. Before we even lift our eye lids, eye movements – such as saccades and microsaccades that let the eyes jump from one to another position – have partially been prepared in the brain stem. Saccades and microsaccades are often assumed to be generated by the same mechanisms. But how saccades and microsaccades can be classified according to shape has not yet been reported in a statistical manner. Research has put more effort into the investigations of microsaccades’ properties and generation only since the last decade. Consequently, we are only beginning to understand the dynamic processes governing microsaccadic eye movements. Within this thesis, the dynamics governing the generation of microsaccades is assessed and the development of a model for the underlying processes. Eye movement trajectories from different experiments are used, recorded with a video-based eye tracking technique, and a novel method is proposed for the scale-invariant detection of saccades (events of large amplitude) and microsaccades (events of small amplitude). Using a time-frequency approach, the method is examined with different experiments and validated against simulated data. A shape model is suggested that allows for a simple estimation of saccade- and microsaccade related properties. For sequences of microsaccades, in this thesis a time-dynamic Markov model is proposed, with a memory horizon that changes over time and which can best describe sequences of microsaccades.
Several authors highlighted that the time course of an experiment itself could have a substantial influence on the interpretability of experimental effects. Since mixed effects modeling had enabled researchers to investigate more complex problems with more precision than before, two naming experiments were conducted with college students, with and without non-words intermixed, and analyzed with regard to frequency, quality, interactive and trial-history effects. The present analyses build on and extend the Bates, Kliegl, Vasishth, and Baayen (2015) approach in order to converge on a parsimonious model that accounts for autocorrelated errors caused by trial history. For three of four cases, a history-sensitive model improved the model fit over a history-naïve model and explained more deviance. In one of these cases, the herein presented approach helped reveal an interaction between stimulus frequency and quality that was not significant without a trial history account. Main and joint effects, limitations, as well as directions for further research, are briefly discussed.
Intuitively, it is clear that neural processes and eye movements in reading are closely connected, but only few studies have investigated both signals simultaneously. Instead, the usual approach is to record them in separate experiments and to subsequently consolidate the results. However, studies using this approach have shown that it is feasible to coregister eye movements and EEG in natural reading and contributed greatly to the understanding of oculomotor processes in reading. The present thesis builds upon that work, assessing to what extent coregistration can be helpful for sentence processing research.
In the first study, we explore how well coregistration is suited to study subtle effects common to psycholinguistic experiments by investigating the effect of distance on dependency resolution. The results demonstrate that researchers must improve the signal-to-noise ratio to uncover more subdued effects in coregistration. In the second study, we compare oscillatory responses in different presentation modes. Using robust effects from world knowledge violations, we show that the generation and retrieval of memory traces may differ between natural reading and word-by-word presentation. In the third study, we bridge the gap between our knowledge of behavioral and neural responses to integration difficulties in reading by analyzing the EEG in the context of regressive saccades. We find the P600, a neural indicator of recovery processes, when readers make a regressive saccade in response to integration difficulties.
The results in the present thesis demonstrate that coregistration can be a useful tool for the study of sentence processing. However, they also show that it may not be suitable for some questions, especially if they involve subtle effects.
MULTILIT
(2015)
This paper presents an overview of the linguistic analyses developed in the MULTILIT project and the processing of the oral and written texts collected. The project investigates the language abilities of multilingual children and adolescents, in particular, those who have Turkish and/or Kurdish as a mother tongue. A further aim of the project is to examine from a psycholinguistic and sociolinguistic perspective the extent to which competence in academic registers is achieved on the basis of the languages spoken by the children, including the language(s) spoken at the home, the language of the country of residence and the first foreign language. To be able to examine these questions using corpus linguistic parameters, we created categories of analysis in MULTILIT.
The data collection comprises texts from bilingual and monolingual children and adolescents in Germany in their first language Turkish, their second language German und their foreign language English. Pupils aged between nine and twenty years of age produced monologue oral and written texts in the two genres of narrative and discursive. On the basis of these samples, we examine linguistic features such as lexical expression (lexical density, lexical diversity), syntactic complexity (syntactic and discursive packaging) as well as phonology in the oral texts and orthography in the written texts, with the aim of investigating the pupils’ growing mastery of these features in academic and informal registers.
To this end the raw data have been transcribed by the use of transcription conventions developed especially for the needs of the MULTILIT data. They are based on the commonly used HIAT and GAT transcription conventions and supplemented with conventions that provide additional information such as features at the graphic level.
The categories of analysis comprise a large number of linguistic categories such as word classes, syntax, noun phrase complexity, complex verbal morphology, direct speech and text structures. We also annotate errors and norm deviations at a wide range of levels (orthographic, morphological, lexical, syntactic and textual). In view of the different language systems, these criteria are considered separately for all languages investigated in the project.
Die Schule steckt mitten in einem Umbruch: Dieser wird durch verschiedene Faktoren wie sinkende Schülerzahlen, Zu- und Abwanderungen von Familien mit ihren Kindern, wachsenden Zahlen von Schulpflichtigen mit Migrationshintergrund und anderes mehr gekennzeichnet. Damit steht die Schule vor neuen Herausforderungen. Außerdem hält der Trend zum Gymnasium an. Auch wechselvolle Dauerreformen schaffen unübersichtliche Schulstrukturen und erschweren sichere Orientierung. Und: Im zunehmend raueren Wettbewerb der Schultypen laufen den öffentlichen Schulen immer mehr Schüler davon und zu den Privatschulen über. Schon jetzt sehen sich zahlreiche Kommunen gezwungen, ihre Bildungsangebote zurückzuschrauben und ihre Schulen zu schließen.
Kann eine Kommune ihren Bürgern aber nicht mehr die schulische „Grundversorgung“ anbieten, so hat dies weitreichende Folgen. Wo Schulen schließen müssen, „stirbt auch der Ort“. Ein umfangreiches, vielfältiges und flächendeckendes Bildungsangebot dagegen ist für die Kommune tragender Pfeiler einer funktionsfähigen Infrastruktur. Hier setzt die 21. Fachtagung des Kommunalwissenschaftlichen Institutes der Universität Potsdam an. Sie behandelt zentrale Themen der nachhaltigen Gewährleistung schulischer Infrastruktur in den Kommunen einschließlich der dazugehörigen Erfahrungsberichte, die über Best-Practice-Modelle sowie über Erfolgsbedingungen und Fallstricke in der Verwaltungspraxis informieren. Damit gibt die Tagung zugleich Impulse den kommunalen Entscheidungsträgern für die Generierung von und den Umgang mit Gestaltungsoptionen zur Standortsicherung im kommunalen Bildungsmanagement.
We consider a Cauchy problem for the heat equation in a cylinder X x (0,T) over a domain X in the n-dimensional space with data on a strip lying on the lateral surface. The strip is of the form
S x (0,T), where S is an open subset of the boundary of X. The problem is ill-posed. Under natural restrictions on the configuration of S we derive an explicit formula for solutions of this problem.
Seit dem UN-Gipfel 1992 in Rio de Janeiro ist die Aufmerksamkeit in Politik und Öffentlichkeit für das Thema „Nachhaltigkeit“ gestiegen. In fast allen Ländern dieser Welt wurden Programme und Maßnahmen zum Schutz und Erhalt der Umwelt und der sozialen Lebensbedingungen umgesetzt. Trotz beachtenswerter Fortschritte sind die bisherigen Effekte jedoch völlig unzureichend. Umso interessanter ist daher der Blick auf einen erfolgreichen Akteur im Bereich der Umwelt- und Nachhaltigkeitspolitik: Kuba. Über diese Erfahrungen des Karibik-Staates wird im deutschen Sprachraum kaum berichtet. Die Autoren leisten hierzu mit ihrer Studie einen Beitrag und analysieren die entsprechenden Politiken, Strategien und Maßnahmen, die in Kuba trotz vielfältiger Probleme zu einer gelungenen Nachhaltigkeitspolitik geführt haben.
Modality in Kakataibo
(2015)
This paper explores the semantic space of modality in Kakataibo
(Panoan). It is found that Kakataibo makes a distinction in the modal
space based on the modality type. Circumstantial modality is encoded
by a construction while the epistemic space is conveyed by the second
position enclitics =dapi ‘inferential’, =id ‘second-hand information’
and =kuni ‘contrastive assertion’. However, none of these strategies to
encode modality restricts the quantificational force, leaving it
underspecified. These facts are consistent with the predictions of
current typologies of modal systems.
It has been observed for many African languages that focussed subjects
have to appear outside of their syntactic base position, as opposed to
focussed objects, which can remain in-situ. This is known as subjectobject
asymmetry of focus marking, which Fiedler et al. (2010) claim
to hold also for Akan. Genzel (2013), on the other hand, argues that
Akan does not exhibit a subject-object focus asymmetry. A questionnaire
study and a production experiment were carried out to investigate
whether focussed subjects may indeed be realized in-situ in Akan. The
results suggest that (i) focussed subjects do not have to be obligatorily
realized ex-situ, and that (ii) the syntactic preference for the realization
of a focussed subject highly depends on exhaustivity.
Sentence type marking is realized by two suffixes in Aymara, one marks
declaratives and the other polar sentences (polar questions and negated
sentences) by picking out one or two propositions, respectively. A third
suffix, initially associated with wh-questions, turns out to be a (scalar)
additive and unrelated to sentence type. The sentence-type-related suffixes
associate with focus and the additive can associate with focus by
attaching to the focused constituent.
According to Aikhenvald (2007:5), descriptive linguistics or linguistic
fieldwork “ideally involves observing the language as it is used,
becoming a member of the community, and often being adopted into
the kinship system”. Descriptive linguistics therefore differs from
theoretical linguistics in that while the former seeks to describe natural
languages as they are used, the latter, other than describing, attempts
to give explanations on how or why language phenomena behave in
certain ways. Thus, I will abstract away from any preconceived ideas
on how sentences ought to be in Awing and take the linguist/reader
through focus and interrogative constructions to get a feeling of how
the Awing people interact verbally.
This paper reopens the discussion on focus marking in Akan (Kwa,
Niger-Congo) by examining the semantics of the so-called focus marker
in the language. It is shown that the so-called focus marker expresses
exhaustivity when it occurs in a sentence with narrow focus. The study
employs four standard tests for exhaustivity proposed in the literature
to examine the semantics of Akan focus constructions (Szabolsci 1981,
1994; É. Kiss 1998; Hartmann and Zimmermann 2007). It is shown that
although a focused entity with the so-called focus marker nà is
interpreted to mean ‘only X and nothing/nobody else,’ this meaning
appears to be pragmatic.
ה"חוק" וה"טבע" בברית המילה
(2015)
בפתח מאמר זה נידון המונח המקראי "חוק" (כבכתוב: אִם בְּחֻקֹּתַי תֵּלֵכוּ". ויקרא כו, ג-ד) באמצעות הצגת נקודות המבט השונות של החכמים הדנים בכך במדרש הארצישראלי "ויקרא רבה".
בחלקו השני של המאמר נידון המקרה הספציפי של מצוות ברית המילה - כדוגמא קלאסית למצווה המכונה בפי החכמים "חוק".
המאמר דן בטעמים השונים שניתנו למצווה זו (כולל אלו שהועלו בתקופה המודרנית ברוח ההסברים האנתרופולוגיים והפסיכואנליטיים); ובסיומו של הדיון מועלה לדיון ויכוח - מקורי עד כמה שניתן לשער - שנשתמר בתלמוד, בין חכם ארצישראלי בן מאה השלישית, רבי עקיבא, ובין רומאי בשם טורנוסרופוס (או טוניוס רופוס) המייצג את תפיסת ה"טבע" הרומית.
האחרון מתקיף את רבי עקיבא בנוגע לברית המילה בטענה שהיהודים מטילים בברית המילה מום בגוף התינוק. תשובת רבי עקיבא מנותחת מנקודת הראות המתמקדת במתח שבין "טבע"
Im Rahmen der EU-weiten REACH-Verordnung haben Alternativmethoden zum Tierversuch in der Toxikologie an Bedeutung gewonnen. Die Alternativmethoden gliedern sich auf in In-vitro- und In-silico-Methoden. In dieser Dissertation wurden verschiedene Konzepte der In-silico-Toxikologie behandelt.
Die bearbeiteten Themen reichen von quantitativen Strukturaktivitätsbeziehungen (QSAR) über eine neue Herangehensweise an das gängige Konzept zur Festlegung von Grenzwerten bis hin zu computerbasierten Modellierungen zum Alkohol- und Bisphenol-A-Stoffwechsel.
Das Kapitel über QSAR befasst sich im Wesentlichen mit der Erstellung und Analyse einer Datenbank mit 878 Substanzen, die sich aus Tierversuchsstudien aus dem Archiv des Bundesinstituts für Risikobewertung zusammensetzt. Das Design wurde dabei an eine bereits bestehende Datenbank angepasst, um so einen möglichst großen Datenpool zu generieren. In der Analyse konnte u.a. gezeigt werden, dass Stoffe mit niedrigerem Molekulargewicht ein erhöhtes Potential für toxikologische Schäden aufwiesen als größere Moleküle.
Mit Hilfe des sogenannten TTC-Konzepts können Grenzwerte für Stoffe geringer Exposition festgelegt werden, zu denen keine toxikologischen Daten zur Verfügung stehen. In dieser Arbeit wurden für die Stoffe dreier Datenbanken entsprechende Grenzwerte festgelegt. Es erfolgte zunächst eine gängige strukturbasierte Aufteilung der Substanzen in die Kategorien "nicht toxisch", "möglicherweise toxisch" und "eindeutig toxisch". Substanzen, die aufgrund ihrer Struktur in eine der drei Klassen eingeordnet werden, erhalten den entsprechenden Grenzwert. Da in die dritte Klasse auch Stoffe eingeordnet werden, deren Toxizität nicht bestimmbar ist, ist sie sehr groß. Daher wurden in dieser Arbeit die ersten beiden Klassen zusammengelgt, um einen größeren Datenpool zu ermöglichen. Eine weitere Neuerung umfasst die Erstellung eines internen Grenzwerts. Diese Vorgehensweise hat den Vorteil, dass der Expositionsweg herausgerechnet wird und somit beispielsweise Studien mit oraler Verabreichung mit Studien dermaler Verabreichung verglichen werden können.
Mittels physiologisch basiertem kinetischem Modelling ist es möglich, Vorgänge im menschlichen Körper mit Hilfe spezieller Software nachzuvollziehen. Durch diese Vorgehensweise können Expositionen von Chemikalien simuliert werden. In einem Teil der Arbeit wurden Alkoholexpositionen von gestillten Neugeborenen simuliert, deren Mütter unmittelbar zuvor alkoholische Getränke konsumiert hatten. Mit dem Modell konnte gezeigt werden, dass die Expositionen des Kindes durchweg gering waren. Nach einem Glas Wein wurden Spitzenkonzentrationen im Blut von Neugeborenen von 0,0034 Promille ermittelt. Zum Vergleich wurde die Exposition durch ein für Säuglinge zugelassenes alkoholhaltiges pflanzliches Arzneimittel simuliert. Hier wurden Spitzenkonzentrationen von 0,0141 Promille erreicht. Daher scheinen Empfehlungen wie gelegentlicher Konsum ohne schädigende Wirkung auf das Kind wissenschaftlich fundiert zu sein.
Ein weiteres Kinetik-Modell befasste sich mit dem Stoffwechsel von Bisphenol A. Teils widersprüchliche Daten zur Belastung mit BPA in der wissenschaftlichen Literatur führen wiederholt zu Anregungen, den Grenzwert der Chemikalie anzupassen. Die Funktionalität der am Metabolismus beteiligten Enzyme kann je nach Individuum unterschiedlich ausgeprägt sein. Mittels Modellings konnte hier gezeigt werden, dass dies maßgeblich dazu führt, dass sich berechnete Plasmaspiegel von Individuen bis zu 4,7-fach unterscheiden.
Die Arbeit konnte somit einen Beitrag zur Nutzung und Weiterentwicklung von In-silico-Modellen für diverse toxikologische Fragestellungen leisten.
Stream water and groundwater are important fresh water resources but their water quality is deteriorated by harmful solutes introduced by human activities. The interface between stream water and the subsurface water is an important zone for retention, transformation and attenuation of these solutes. Streambed structures enhance these processes by increased water and solute exchange across this interface, denoted as hyporheic exchange.
This thesis investigates the influence of hydrological and morphological factors on hyporheic water and solute exchange as well as redox-reactions in fluvial streambed structures on the intermediate scale (10–30m). For this purpose, a three-dimensional numerical modeling approach for coupling stream water flow with porous media flow is used. Multiple steady state stream water flow scenarios over different generic pool-riffle morphologies and a natural in-stream gravel bar are simulated by a computational fluid dynamics code that provides the hydraulic head distribution at the streambed. These heads are subsequently used as the top boundary condition of a reactive transport groundwater model of the subsurface beneath the streambed. Ambient groundwater that naturally interacts with the stream water is considered in scenarios of different magnitudes of downwelling stream water (losing case) and upwelling groundwater (gaining case). Also, the neutral case, where stream stage and groundwater levels are balanced is considered. Transport of oxygen, nitrate and dissolved organic carbon and their reaction by aerobic respiration and denitrification are modeled.
The results show that stream stage and discharge primarily induce hyporheic exchange flux and solute transport with implications for specific residence times and reactions at both the fully and partially submerged structures. Gaining and losing conditions significantly diminish the extent of the hyporheic zone, the water exchange flux, and shorten residence times for both the fully and partially submerged structures. With increasing magnitude of gaining or losing conditions, these metrics exponentially decrease.
Stream water solutes are transported mainly advectively into the hyporheic zone and hence their influx corresponds directly to the infiltrating water flux. Aerobic respiration takes place in the shallow streambed sediments, coinciding to large parts with the extent of the hyporheic exchange flow. Denitrification occurs mainly as a “reactive fringe” surrounding the aerobic zone, where oxygen concentration is low and still a sufficient amount of stream water carbon source is available. The solute consumption rates and the efficiency of the aerobic and anaerobic reactions depend primarily on the available reactive areas and the residence times, which are both controlled by the interplay between hydraulic head distribution at the streambed and the gradients between stream stage and ambient groundwater. Highest solute consumption rates can be expected under neutral conditions, where highest solute flux, longest residence times and largest extent of the hyporheic exchange occur. The results of this thesis show that streambed structures on the intermediate scale have a significant potential to contribute to a net solute turnover that can support a healthy status of the aquatic ecosystem.
Wie verhandelt die Praxis?
(2015)
Aus dem Inhalt:
- 10 Jahre Responsibility to Protect: Ein Sieg für die Menschenrechte? – Eine politik- und rechtswissenschaftliche Analyse
- Neue Regeln zur Abwesenheit des Angeklagten vor dem IStGH:
Menschenrechtliche Anforderungen an In-absentia-Verfahren
- EGMR: S.A.S. ./. Frankreich – Urteilsbesprechung zum Burkaverbot
We study segregation of the subducted oceanic crust (OC) at the core mantle boundary and its ability to accumulate and form large thermochemical piles (such as the seismically observed Large Low Shear Velocity Provinces - LLSVPs). Our high-resolution numerical simulations suggest that the longevity of LLSVPs for up to three billion years, and possibly longer, can be ensured by a balance in the rate of segregation of high-density OC-material to the CMB, and the rate of its entrainment away from the CMB by mantle upwellings.
For a range of parameters tested in this study, a large-scale compositional anomaly forms at the CMB, similar in shape and size to the LLSVPs. Neutrally buoyant thermochemical piles formed by mechanical stirring - where thermally induced negative density anomaly is balanced by the presence of a fraction of dense anomalous material - best resemble the geometry of LLSVPs. Such neutrally buoyant piles tend to emerge and survive for at least 3Gyr in simulations with quite different parameters. We conclude that for a plausible range of values of density anomaly of OC material in the lower mantle - it is likely that it segregates to the CMB, gets mechanically mixed with the ambient material, and forms neutrally buoyant large scale compositional anomalies similar in shape to the LLSVPs.
We have developed an efficient FEM code with dynamically adaptive time and space resolution, and marker-in-cell methodology. This enabled us to model thermochemical mantle convection at realistically high convective vigor, strong thermally induced viscosity variations, and long term evolution of compositional fields.
In this thesis we study reciprocal classes of Markov chains. Given a continuous time Markov chain on a countable state space, acting as reference dynamics, the associated reciprocal class is the set of all probability measures on path space that can be written as a mixture of its bridges. These processes possess a conditional independence property that generalizes the Markov property, and evolved from an idea of Schrödinger, who wanted to obtain a probabilistic interpretation of quantum mechanics.
Associated to a reciprocal class is a set of reciprocal characteristics, which are space-time functions that determine the reciprocal class. We compute explicitly these characteristics, and divide them into two main families: arc characteristics and cycle characteristics. As a byproduct, we obtain an explicit criterion to check when two different Markov chains share their bridges.
Starting from the characteristics we offer two different descriptions of the reciprocal class, including its non-Markov probabilities.
The first one is based on a pathwise approach and the second one on short time asymptotic. With the first approach one produces a family of functional equations whose only solutions are precisely the elements of the reciprocal class. These equations are integration by parts on path space associated with derivative operators which perturb the paths by mean of the addition of random loops. Several geometrical tools are employed to construct such formulas. The problem of obtaining sharp characterizations is also considered, showing some interesting connections with discrete geometry. Examples of such formulas are given in the framework of counting processes and random walks on Abelian groups, where the set of loops has a group structure.
In addition to this global description, we propose a second approach by looking at the short time behavior of a reciprocal process. In the same way as the Markov property and short time expansions of transition probabilities characterize Markov chains, we show that a reciprocal class is characterized by imposing the reciprocal property and two families of short time expansions for the bridges. Such local approach is suitable to study reciprocal processes on general countable graphs. As application of our characterization, we considered several interesting graphs, such as lattices, planar
graphs, the complete graph, and the hypercube.
Finally, we obtain some first results about concentration of measure implied by lower bounds on the reciprocal characteristics.
The relationship between nutrition and the development of chronic diseases including metabolic syndrome, diabetes mellitus, cancer and cardiovascular disease has been well studied. On the other hand, changes in the GH-IGF-1 axis in association with nutrition-related diseases have been reported. The interplay between GH, total IGF-1 and different inhibitory and stimulatory kinds of IGF-1 binding proteins (IGFBPs) results in IGF-1 bioactivity, the ability of IGF-1 to induce phosphorylation of its receptor and consequently its signaling. Moreover, IGF-1 bioactivity is sufficient to reflect any change in the GH-IGF-1 system. Accumulating evidence suggests that both of high protein diet, characterized by increased glucagon secretion, and insulin-induced hypoglycemia increase mortality rate and the mechanisms are unclear. However both of glucagon and insulin-induced hypoglycemia are potent stimuli of GH secretion. The aim of the current study was to identify the impact of glucagon and insulin-induced hypoglycemia on IGF-1 bioactivity as possible mechanisms. In a double-blind placebo-controlled study, glucagon was intramuscularly administrated in 13 type 1 diabetic patients (6 males /7 females; [BMI]: 24.8 ± 0.95 kg/m2), 11 obese subjects (OP; 5/ 6; 34.4 ± 1.7 kg/m2), and 13 healthy lean participants (LP; 6/ 7; 21.7 ± 0.6 kg/m2), whereas 12 obese subjects (OP; 6/ 6; 34.4 ± 1.7 kg/m2), and 13 healthy lean participants (LP; 6/ 7; 21.7 ± 0.6 kg/m2) performed insulin tolerance test in another double-blind placebo-controlled study and changes in GH, total IGF-1, IGF binding proteins (IGFBPs) and IGF-1 bioactivity, measured by the cell-based KIRA method, were investigated. In addition, the interaction between the metabolic hormones (glucagon and insulin) and the GH-IGF-1 system on the transcriptional level was studied using mouse primary hepatocytes. In this thesis, glucagon decreased IGF-1 bioactivity in humans independently of endogenous insulin levels, most likely through modulation of IGFBP-1 and-2 levels. The glucagon-induced reduction in IGF-1 bioactivity may represent a novel mechanism underlying the impact of glucagon on GH secretion and may explain the negative effect of high protein diet related to increased cardiovascular risk and mortality rate. In addition, insulin-induced hypoglycemia was correlated with a decrease in IGF-1 bioactivity through up-regulation of IGFBP-2. These results may refer to a possible and poorly explored mechanism explaining the strong association between hypoglycemia and increased cardiovascular mortality among diabetic patients.
This dissertation investigates the working memory mechanism subserving human sentence processing and its relative contribution to processing difficulty as compared to syntactic prediction. Within the last decades, evidence for a content-addressable memory system underlying human cognition in general has accumulated (e.g., Anderson et al., 2004). In sentence processing research, it has been proposed that this general content-addressable architecture is also used for language processing (e.g., McElree, 2000).
Although there is a growing body of evidence from various kinds of linguistic dependencies that is consistent with a general content-addressable memory subserving sentence processing (e.g., McElree et al., 2003; VanDyke2006), the case of reflexive-antecedent dependencies has challenged this view. It has been proposed that in the processing of reflexive-antecedent dependencies, a syntactic-structure based memory access is used rather than cue-based retrieval within a content-addressable framework (e.g., Sturt, 2003).
Two eye-tracking experiments on Chinese reflexives were designed to tease apart accounts assuming a syntactic-structure based memory access mechanism from cue-based retrieval (implemented in ACT-R as proposed by Lewis and Vasishth (2005).
In both experiments, interference effects were observed from noun phrases which syntactically do not qualify as the reflexive's antecedent but match the animacy requirement the reflexive imposes on its antecedent. These results are interpreted as evidence against a purely syntactic-structure based memory access. However, the exact pattern of effects observed in the data is only partially compatible with the Lewis and Vasishth cue-based parsing model.
Therefore, an extension of the Lewis and Vasishth model is proposed. Two principles are added to the original model, namely 'cue confusion' and 'distractor prominence'.
Although interference effects are generally interpreted in favor of a content-addressable memory architecture, an alternative explanation for interference effects in reflexive processing has been proposed which, crucially, might reconcile interference effects with a structure-based account.
It has been argued that interference effects do not necessarily reflect cue-based retrieval interference in a content-addressable memory but might equally well be accounted for by interference effects which have already occurred at the moment of encoding the antecedent in memory (Dillon, 2011).
Three experiments (eye-tracking and self-paced reading) on German reflexives and Swedish possessives were designed to tease apart cue-based retrieval interference from encoding interference. The results of all three experiments suggest that there is no evidence that encoding interference affects the retrieval of a reflexive's antecedent.
Taken together, these findings suggest that the processing of reflexives can be explained with the same cue-based retrieval mechanism that has been invoked to explain syntactic dependency resolution in a range of other structures. This supports the view that the language processing system is located within a general cognitive architecture, with a general-purpose content-addressable working memory system operating on linguistic expressions.
Finally, two experiments (self-paced reading and eye-tracking) using Chinese relative clauses were conducted to determine the relative contribution to sentence processing difficulty of working-memory processes as compared to syntactic prediction during incremental parsing.
Chinese has the cross-linguistically rare property of being a language with subject-verb-object word order and pre-nominal relative clauses. This property leads to opposing predictions of expectation-based
accounts and memory-based accounts with respect to the relative processing difficulty of subject vs. object relatives.
Previous studies showed contradictory results, which has been attributed to different kinds local ambiguities confounding the materials (Lin and Bever, 2011). The two experiments presented are the first to compare Chinese relatives clauses in syntactically unambiguous contexts.
The results of both experiments were consistent with the predictions of the expectation-based account of sentence processing but not with the memory-based account. From these findings, I conclude that any theory of human sentence processing needs to take into account the power of predictive processes unfolding in the human mind.
A lot has been published about the competencies needed by
students in the 21st century (Ravenscroft et al., 2012). However, equally
important are the competencies needed by educators in the new era
of digital education. We review the key competencies for educators in
light of the new methods of teaching and learning proposed by Massive
Open Online Courses (MOOCs) and their on-campus counterparts,
Small Private Online Courses (SPOCs).
Participants of this workshop will be confronted exemplarily
with a considerable inconsistency of global Informatics education at
lower secondary level. More importantly, they are invited to contribute
actively on this issue in form of short case studies of their countries.
Until now, very few countries have been successful in implementing
Informatics or Computing at primary and lower secondary level. The
spectrum from digital literacy to informatics, particularly as a discipline
in its own right, has not really achieved a breakthrough and seems to
be underrepresented for these age groups. The goal of this workshop
is not only to discuss the anamnesis and diagnosis of this fragmented
field, but also to discuss and suggest viable forms of therapy in form of
setting educational standards. Making visible good practices in some
countries and comparing successful approaches are rewarding tasks for
this workshop.
Discussing and defining common educational standards on a transcontinental
level for the age group of 14 to 15 years old students in a readable,
assessable and acceptable form should keep the participants of this
workshop active beyond the limited time at the workshop.
Let’s talk about CS!
(2015)
To communicate about a science is the most important key
competence in education for any science. Without communication we
cannot teach, so teachers should reflect about the language they use in
class properly. But the language students and teachers use to communicate
about their CS courses is very heterogeneous, inconsistent and
deeply influenced by tool names. There is a big lack of research and
discussion in CS education regarding the terminology and the role of
concepts and tools in our science. We don’t have a consistent set of
terminology that we agree on to be helpful for learning our science.
This makes it nearly impossible to do research on CS competencies as
long as we have not agreed on the names we use to describe these. This
workshop intends to provide room to fill with discussion and first ideas
for future research in this field.
ProtoSense
(2015)
The poster and abstract describe the importance of teaching
information security in school. After a short description of information
security and important aspects, I will show, how information security
fits into different guidelines or models for computer science educations
and that it is therefore on of the key competencies. Afterwards I will
present you a rough insight of teaching information security in Austria.
Current curricular trends require teachers in Baden-
Wuerttemberg (Germany) to integrate Computer Science (CS) into
traditional subjects, such as Physical Science. However, concrete guidelines
are missing. To fill this gap, we outline an approach where a
microcontroller is used to perform and evaluate measurements in the
Physical Science classroom.
Using the open-source Arduino platform, we expect students to acquire
and develop both CS and Physical Science competencies by using a
self-programmed microcontroller. In addition to this combined development
of competencies in Physical Science and CS, the subject matter
will be embedded in suitable contexts and learning environments,
such as weather and climate.
Think logarithmically!
(2015)
We discuss here a number of algorithmic topics which we
use in our teaching and in learning of mathematics and informatics to
illustrate and document the power of logarithm in designing very efficient
algorithms and computations – logarithmic thinking is one of the
most important key competencies for solving real world practical problems.
We demonstrate also how to introduce logarithm independently
of mathematical formalism using a conceptual model for reducing a
problem size by at least half. It is quite surprising that the idea, which
leads to logarithm, is present in Euclid’s algorithm described almost
2000 years before John Napier invented logarithm.
A project involving the composition of a number of pieces
of music by public participants revealed levels of engagement with and
mastery of complex music technologies by a number of secondary student
volunteers. This paper reports briefly on some initial findings of
that project and seeks to illuminate an understanding of computational
thinking across the curriculum.
Physical computing covers the design and realization of interactive
objects and installations and allows students to develop concrete,
tangible products of the real world that arise from the learners’
imagination. This way, constructionist learning is raised to a level that
enables students to gain haptic experience and thereby concretizes the
virtual. In this paper the defining characteristics of physical computing
are described. Key competences to be gained with physical computing
will be identified.
Mentoring in a Digital World
(2015)
This paper focuses on the results of the evaluation of the first
pilot of an e-mentoring unit designed by the Hands-On ICT consortium,
funded by the EU LLL programme. The overall aim of this two-year
activity is to investigate the value for professional learning of Massive
Online Open Courses (MOOCs) and Community Online Open Courses
(COOCs) in the context of a ‘community of practice’. Three units in the
first pilot covered aspects of using digital technologies to develop creative
thinking skills. The findings in this paper relate to the fourth unit
about e-mentoring, a skill that was important to delivering the course
content in the other three units. Findings about the e-mentoring unit
included: the students’ request for detailed profiles so that participants
can get to know each other; and, the need to reconcile the different
interpretations of e-mentoring held by the participants when the course
begins. The evaluators concluded that the major issues were that: not all
professional learners would self-organise and network; and few would
wish to mentor their colleagues voluntarily. Therefore, the e-mentoring
issues will need careful consideration in pilots two and three to identify
how e-mentoring will be organised.
The study reported in this paper involved the employment
of specific in-class exercises using a Personal Response System (PRS).
These exercises were designed with two goals: to enhance students’
capabilities of tracing a given code and of explaining a given code in
natural language with some abstraction. The paper presents evidence
from the actual use of the PRS along with students’ subjective impressions
regarding both the use of the PRS and the special exercises. The
conclusions from the findings are followed with a short discussion on
benefits of PRS-based mental processing exercises for learning programming
and beyond.
In this paper we describe the recent state of our research
project concerning computer science teachers’ knowledge on students’
cognition. We did a comprehensive analysis of textbooks, curricula
and other resources, which give teachers guidance to formulate assignments.
In comparison to other subjects there are only a few concepts
and strategies taught to prospective computer science teachers in university.
We summarize them and given an overview on our empirical
approach to measure this knowledge.
How does the Implementation of a Literacy Learning Tool Kit influence Literacy Skill Acquisition?
(2015)
This study aimed at following how teachers transfer skills
into results while using ABRA literacy software. This was done in
the second part of the pilot study whose aim was to provide equity to
control group teachers and students by exposing them to the ABRACADABRA
treatment after the end of phase 1. This opportunity was
used to follow the phase 1 teachers to see how the skills learned were
being transformed into results. A standard three-day initial training and
planning session on how to use ABRA to teach literacy was held at the
beginning of each phase for ABRA teachers (phase 1 experimental and
phase 2 delayed ABRA). Teachers were provided with teaching materials
including a tentative ABRA curriculum developed to align with the
Kenyan English Language requirements for year 1 and 3 students. Results
showed that although there was no significant difference between
the groups in vocabulary-related subscales which include word reading
and meaning as well as sentence comprehension, students in ABRACADABRA
classes improved their scores at a significantly higher rate
than students in control classes in comprehension related scores. An
average student in the ABRACADABRA group improved by 12 and
16 percentile points respectively compared to their counterparts in the
control group.
The Technology Proficiency Self-Assessment (TPSA) questionnaire
has been used for 15 years in the USA and other nations as a
self-efficacy measure for proficiencies fundamental to effective technology
integration in the classroom learning environment. Internal consistency
reliabilities for each of the five-item scales have typically ranged
from .73 to .88 for preservice or inservice technology-using teachers.
Due to changing technologies used in education, researchers sought to
renovate partially obsolete items and extend self-efficacy assessment to
new areas, such as social media and mobile learning. Analysis of 2014
data gathered on a new, 34 item version of the TPSA indicates that the
four established areas of email, World Wide Web (WWW), integrated
applications, and teaching with technology continue to form consistent
scales with reliabilities ranging from .81 to .93, while the 14 new items
gathered to represent emerging technologies and media separate into
two scales, each with internal consistency reliabilities greater than .9.
The renovated TPSA is deemed to be worthy of continued use in the
teaching with technology context.
Computational Thinking
(2015)
Digital technology has radically changed the way people
work in industry, finance, services, media and commerce. Informatics
has contributed to the scientific and technological development of our
society in general and to the digital revolution in particular. Computational
thinking is the term indicating the key ideas of this discipline that
might be included in the key competencies underlying the curriculum
of compulsory education. The educational potential of informatics has
a history dating back to the sixties. In this article, we briefly revisit this
history looking for lessons learned. In particular, we focus on experiences
of teaching and learning programming. However, computational
thinking is more than coding. It is a way of thinking and practicing interactive
dynamic modeling with computers. We advocate that learners
can practice computational thinking in playful contexts where they can
develop personal projects, for example building videogames and/or robots,
share and discuss their construction with others. In our view, this
approach allows an integration of computational thinking in the K-12
curriculum across disciplines.
How Things Work
(2015)
Recognizing and defining functionality is a key competence
adopted in all kinds of programming projects. This study investigates
how far students without specific informatics training are able to identify
and verbalize functions and parameters. It presents observations
from classroom activities on functional modeling in high school chemistry
lessons with altogether 154 students. Finally it discusses the potential
of functional modelling to improve the comprehension of scientific
content.
This paper originated from discussions about the need for
important changes in the curriculum for Computing including two focus
group meetings at IFIP conferences over the last two years. The
paper examines how recent developments in curriculum, together with
insights from curriculum thinking in other subject areas, especially mathematics
and science, can inform curriculum design for Computing.
The analysis presented in the paper provides insights into the complexity
of curriculum design as well as identifying important constraints and
considerations for the ongoing development of a vision and framework
for a Computing curriculum.
This article shows a discussion about the key competencies
in informatics and ICT viewed from a philosophical foundation presented
by Martha Nussbaum, which is known as ‘ten central capabilities’.
Firstly, the outline of ‘The Capability Approach’, which has been presented
by Amartya Sen and Nussbaum as a theoretical framework of
assessing the state of social welfare, will be explained. Secondly, the
body of Nussbaum’s ten central capabilities and the reason for being
applied as the basis of discussion will be shown. Thirdly, the relationship
between the concept of ‘capability’ and ‘competency’ is to be
discussed. After that, the author’s assumption of the key competencies
in informatics and ICT led from the examination of Nussbaum’s ten
capabilities will be presented.
The objectives of this study were to examine (a) the effect
of dynamic assessment (DA) in a 3D Immersive Virtual Reality
(IVR) environment as compared with computerized 2D and noncomputerized
(NC) situations on cognitive modifiability, and (b) the
transfer effects of these conditions on more difficult problem solving
administered two weeks later in a non-computerized environment. A
sample of 117 children aged 6:6-9:0 years were randomly assigned
into three experimental groups of DA conditions: 3D, 2D, and NC, and
one control group (C). All groups received the pre- and post-teaching
Analogies subtest of the Cognitive Modifiability Battery (CMB-AN).
The experimental groups received a teaching phase in conditions similar
to the pre-and post-teaching phases. The findings showed that cognitive
modifiability, in a 3D IVR, was distinctively higher than in the two
other experimental groups (2D computer group and NC group). It was
also found that the 3D group showed significantly higher performance
in transfer problems than the 2D and NC groups.
BugHunt
(2015)
Competencies related to operating systems and computer
security are usually taught systematically. In this paper we present
a different approach, in which students have to remove virus-like
behaviour on their respective computers, which has been induced by
software developed for this purpose. They have to develop appropriate
problem-solving strategies and thereby explore essential elements of
the operating system. The approach was implemented exemplarily in
two computer science courses at a regional general upper secondary
school and showed great motivation and interest in the participating
students.
In the project MoKoM, which is funded by the German
Research Foundation (DFG) from 2008 to 2012, a test instrument
measuring students’ competences in computer science was developed.
This paper presents the results of an expert rating of the levels of
students’ competences done for the items of the instrument.
At first we will describe the difficulty-relevant features that were
used for the evaluation. These were deduced from computer science,
psychological and didactical findings and resources. Potentials and
desiderata of this research method are discussed further on. Finally
we will present our conclusions on the results and give an outlook on
further steps.
The growing impact of globalisation and the development of
a ‘knowledge society’ have led many to argue that 21st century skills are
essential for life in twenty-first century society and that ICT is central
to their development. This paper describes how 21st century skills, in
particular digital literacy, critical thinking, creativity, communication
and collaboration skills, have been conceptualised and embedded in the
resources developed for teachers in iTEC, a four-year, European project.
The effectiveness of this approach is considered in light of the data
collected through the evaluation of the pilots, which considers both the
potential benefits of using technology to support the development of
21st century skills, but also the challenges of doing so. Finally, the paper
discusses the learning support systems required in order to transform
pedagogies and embed 21st century skills. It is argued that support is
required in standards and assessment; curriculum and instruction; professional
development; and learning environments.
This paper discusses results from a small-scale research
study, together with some recently published research into student
perceptions of ICT for learning in schools, to consider relevant skills
that do not appear to currently being taught. The paper concludes by
raising three issues relating to learning with and through ICT that need
to be addressed in school curricula and classroom teaching.
The Student Learning Ecology
(2015)
Educational research on social media has showed that
students use it for socialisation, personal communication, and informal
learning. Recent studies have argued that students to some degree use
social media to carry out formal schoolwork. This article gives an
explorative account on how a small sample of Norwegian high school
students use social media to self-organise formal schoolwork. This
user pattern can be called a “student learning ecology”, which is a
user perspective on how participating students gain access to learning
resources.
Teaching Data Management
(2015)
Data management is a central topic in computer science as
well as in computer science education. Within the last years, this topic is
changing tremendously, as its impact on daily life becomes increasingly
visible. Nowadays, everyone not only needs to manage data of various
kinds, but also continuously generates large amounts of data. In
addition, Big Data and data analysis are intensively discussed in public
dialogue because of their influences on society. For the understanding of
such discussions and for being able to participate in them, fundamental
knowledge on data management is necessary. Especially, being aware
of the threats accompanying the ability to analyze large amounts of
data in nearly real-time becomes increasingly important. This raises the
question, which key competencies are necessary for daily dealings with
data and data management.
In this paper, we will first point out the importance of data management
and of Big Data in daily life. On this basis, we will analyze which are
the key competencies everyone needs concerning data management to
be able to handle data in a proper way in daily life. Afterwards, we will
discuss the impact of these changes in data management on computer
science education and in particular database education.
Social networks are currently at the forefront of tools that
lend to Personal Learning Environments (PLEs). This study aimed to
observe how students perceived PLEs, what they believed were the
integral components of social presence when using Facebook as part
of a PLE, and to describe student’s preferences for types of interactions
when using Facebook as part of their PLE. This study used mixed
methods to analyze the perceptions of graduate and undergraduate
students on the use of social networks, more specifically Facebook as a
learning tool. Fifty surveys were returned representing a 65 % response
rate. Survey questions included both closed and open-ended questions.
Findings suggested that even though students rated themselves relatively
well in having requisite technology skills, and 94 % of students used
Facebook primarily for social use, they were hesitant to migrate these
skills to academic use because of concerns of privacy, believing that
other platforms could fulfil the same purpose, and by not seeing the
validity to use Facebook in establishing social presence. What lies
at odds with these beliefs is that when asked to identify strategies in
Facebook that enabled social presence to occur in academic work, the
majority of students identified strategies in five categories that lead to
social presence establishment on Facebook during their coursework.
The paper discusses the issue of supporting informatics
(computer science) education through competitions for lower and
upper secondary school students (8–19 years old). Competitions play
an important role for learners as a source of inspiration, innovation,
and attraction. Running contests in informatics for school students
for many years, we have noticed that the students consider the contest
experience very engaging and exciting as well as a learning experience.
A contest is an excellent instrument to involve students in problem
solving activities. An overview of infrastructure and development
of an informatics contest from international level to the national one
(the Bebras contest on informatics and computer fluency, originated
in Lithuania) is presented. The performance of Bebras contests in 23
countries during the last 10 years showed an unexpected and unusually
high acceptance by school students and teachers. Many thousands of
students participated and got a valuable input in addition to their regular
informatics lectures at school. In the paper, the main attention is paid
to the developed tasks and analysis of students’ task solving results in
Lithuania.
The paper presents two approaches to the development of
a Computer Science Competence Model for the needs of curriculum
development and evaluation in Higher Education. A normativetheoretical
approach is based on the AKT and ACM/IEEE curriculum
and will be used within the recommendations of the German
Informatics Society (GI) for the design of CS curricula. An empirically
oriented approach refines the categories of the first one with regard to
specific subject areas by conducting content analysis on CS curricula of
important universities from several countries. The refined model will be
used for the needs of students’ e-assessment and subsequent affirmative
action of the CS departments.
Regardless of what is intended by government curriculum
specifications and advised by educational experts, the competencies
taught and learned in and out of classrooms can vary considerably.
In this paper, we discuss in particular how we can investigate the
perceptions that individual teachers have of competencies in ICT,
and how these and other factors may influence students’ learning. We
report case study research which identifies contradictions within the
teaching of ICT competencies as an activity system, highlighting issues
concerning the object of the curriculum, the roles of the participants and
the school cultures. In a particular case, contradictions in the learning
objectives between higher order skills and the use of application tools
have been resolved by a change in the teacher’s perceptions which
have not led to changes in other aspects of the activity system. We look
forward to further investigation of the effects of these contradictions in
other case studies and on forthcoming curriculum change.
As a result of the Bologna reform of educational systems in
Europe the outcome orientation of learning processes, competence-oriented
descriptions of the curricula and competence-oriented assessment
procedures became standard also in Computer Science Education
(CSE). The following keynote addresses important issues of shaping
a CSE competence model especially in the area of informatics system
comprehension and object-oriented modelling. Objectives and research
methodology of the project MoKoM (Modelling and Measurement
of Competences in CSE) are explained. Firstly, the CSE competence
model was derived based on theoretical concepts and then secondly the
model was empirically examined and refined using expert interviews.
Furthermore, the paper depicts the development and examination of
a competence measurement instrument, which was derived from the
competence model. Therefore, the instrument was applied to a large
sample of students at the gymnasium’s upper class level. Subsequently,
efforts to develop a competence level model, based on the retrieved empirical
results and on expert ratings are presented. Finally, further demands
on research on competence modelling in CSE will be outlined.
Computational thinking is a fundamental skill set that is learned
by studying Informatics and ICT. We argue that its core ideas can
be introduced in an inspiring and integrated way to both teachers and
students using fun and contextually rich cs4fn ‘Computer Science for
Fun’ stories combined with ‘unplugged’ activities including games and
magic tricks. We also argue that understanding people is an important
part of computational thinking. Computational thinking can be fun for
everyone when taught in kinaesthetic ways away from technology.
Graph databases provide a natural way of storing and querying graph data. In contrast to relational databases, queries over graph databases enable to refer directly to the graph structure of such graph data. For example, graph pattern matching can be employed to formulate queries over graph data.
However, as for relational databases running complex queries can be very time-consuming and ruin the interactivity with the database. One possible approach to deal with this performance issue is to employ database views that consist of pre-computed answers to common and often stated queries. But to ensure that database views yield consistent query results in comparison with the data from which they are derived, these database views must be updated before queries make use of these database views. Such a maintenance of database views must be performed efficiently, otherwise the effort to create and maintain views may not pay off in comparison to processing the queries directly on the data from which the database views are derived.
At the time of writing, graph databases do not support database views and are limited to graph indexes that index nodes and edges of the graph data for fast query evaluation, but do not enable to maintain pre-computed answers of complex queries over graph data. Moreover, the maintenance of database views in graph databases becomes even more challenging when negation and recursion have to be supported as in deductive relational databases.
In this technical report, we present an approach for the efficient and scalable incremental graph view maintenance for deductive graph databases. The main concept of our approach is a generalized discrimination network that enables to model nested graph conditions including negative application conditions and recursion, which specify the content of graph views derived from graph data stored by graph databases. The discrimination network enables to automatically derive generic maintenance rules using graph transformations for maintaining graph views in case the graph data from which the graph views are derived change. We evaluate our approach in terms of a case study using multiple data sets derived from open source projects.
Ziel dieser Arbeit war die Synthese und Charakterisierung von neuartigen fluoreszierenden Copolymeren zur Analytdetektion in wässrigen Systemen. Das Detektionssystem sollte ein einfaches Schalten der Fluoreszenz bei Analytbindung „Aus“ bzw. Verdrängung „An“ ermöglichen. Dafür wurde die Synthese eines funktionalisierten Monomers so geplant, dass sich Fluorophor und Analyt innerhalb derselben Monomereinheit in direkter Nachbarschaft zueinander befinden. So sollten bei Erkennung des Analyten durch eine mit einem Fluoreszenzlöscher funktionalisierte Erkennungsstruktur Fluorophor und Löscher in einen vorgegebenen Abstand zueinander gezwungen und die Fluoreszenz des Fluorophors effizient gelöscht werden. Bei anschließender Verdrängung der Erkennungseinheit durch einen stärker bindenden Analyten sollte die Fluoreszenz wieder „angeschaltet“ werden. Eine weitere Zielstellung für das Detektionssystem war eine hohe Löslichkeit und Fluoreszenzintensität in Wasser. Da die Anwendung solcher Sensoren besonders in der Medizin und Biologie, z.B. für Schnellerkennungstest von Pathogenen, von Interesse ist, ist die Kompatibilität mit wässrigen Medien essentiell. Die funktionalisierten Monomere wurden frei radikalisch mit N Vinyl-pyrrolidon bzw. N Vinyl¬caprolactam zu wasserlöslichen, fluoreszierenden Copolymeren umgesetzt. In den N-Vinyl¬pyrrolidon-Polymeren (PNVP) wurde RhodaminB, in den thermoresponsiven N Vinyl¬caprolactam-Polymeren (PNVCL) ein Naphthalsäureimid als Fluorophor verwendet. Während Rhodamine eine hohe Fluoreszenzintensität, gute Quantenausbeuten und hohen Extinktionskoeffizienten in Wasser zeigen, sind Naphthalsäure¬imide umgebungssensitive Chromophore, die bei Änderung ihrer Lösungsmittelumgebung, wie z.B. beim Kollaps eines thermoresponsiven Polymers in Wasser, ihre Fluoreszenzintensität und Quantenausbeute drastisch ändern können. Der Vorteil der hier verwendeten Strategie der Monomersynthese liegt darin, dass bei jeder spezifischen Analytdetektion durch eine Erkennungseinheit die Fluoreszenz effizient gelöscht bzw. bei Verdrängung durch einen stärker bindenden Analyten wieder „angeschaltet“ wird. Dieses Prinzip wird bereits vielfach in der Biologie in sogenannten „Molecular Beacons“ ausgenutzt, wobei ein Fluorophor und ein Löscher durch spezifische DNA Basenpaarung in einen vorgegebenen Abstand zueinander gezwungen werden und so ein „Schalten“ der Fluoreszenz ermöglichen. Aufgrund der vorgegebenen Struktur der DNA Basensequenzen ist es jedoch nicht direkt auf andere Erkennungsreaktionen übertragbar. Daher wurde ein Modellsystem entwickelt, welches die Möglichkeit bietet Analyt, Erkennungseinheit und Signalgeber variabel, je nach Anforderungen des Systems, auszutauschen. So soll es möglich sein, den Sensor a priori für jede Erkennungs¬reaktion zu verwenden. Als Modell Bindungs¬paare wurden ß Cyclodextrin/Adamantan und Con¬cana¬valinA/Mannose ausgewählt. Adamantan bzw. Mannose wurde als Analyt zusammen mit dem Fluorophor in das Polymer eingebunden. ß Cyclo¬dextrin (ß CD) bzw. ConcanavalinA (ConA) wurde als Erkennungsstruktur an einem Fluoreszenzlöscher immobilisiert. Polymer-basierte Fluoreszenzsensoren sind in der Fachliteratur gut dokumentiert. In der Regel sind Signalgeber und Analyt jedoch statistisch im Polymer verteilt, da sie sich entweder in unterschiedlichen Monomereinheiten befinden oder die Funktionalisierung durch eine polymeranaloge Umsetzung erfolgt. Der gewählte Ansatz Fluorophor und Analyt innerhalb derselben Monomereinheit einzubinden, soll bei jeder Erkennungsreaktion des Analyten zu einer Änderung der Signalintensität des Fluorophors führen. Eine hohe Signalintensität bei Analytdetektion ist wünschenswert, insbesondere für Erkennungsreaktionen, die mit möglichst geringem apparativem Aufwand, am besten mit dem bloßen Auge zu verfolgen sein sollen. Des Weiteren ist es möglich den Fluorophorgehalt im Polymer genau einzustellen und so Selbstlöschung zu vermeiden. Die synthetisierten Polymere haben einen Fluorophorgehalt von 0,01 mol% bis 0,5 mol%. Für die RhodaminB haltigen Polymere zeigte sich, dass ein Fluorophorgehalt unterhalb 0,1 mol% im Polymer die höchsten Ausbeuten, Molmassen und Quantenausbeuten liefert. Für die Naphthalsäureimid haltigen Polymere hingegen wurden auch für einen Fluorophorgehalt von bis zu 1 mol% hohe Ausbeuten und Molmassen erreicht. Die Naphthalsäureimid haltigen Polymere haben jedoch in wässriger Lösungsmittelumgebung nur geringe Quantenausbeuten. Als Fluoreszenzlöscher wurden Goldnanopartikel synthetisiert, die mit den entsprechenden Erkennungsstrukturen (ß-CD oder ConA) für den verwendeten Analyten funktionalisiert wurden. Goldnanopartikel als Löscher bieten den Vorteil, dass ihre Dispergierbarkeit in einem Lösemittel durch Funktionalisierung ihrer Hülle gezielt gesteuert werden kann. Durch die hohe Affinität von Goldnanopartikeln zu Thiolen und Aminen konnten sie mit Hilfe einfacher Syntheseschritte mit Thio ß CD Derivaten bzw. ConA funktionalisiert werden. In der hier vorgelegten Arbeit sollte ein Modellsystem für einen solches fluoreszenz-basiertes Detektionssystem in Wasser entwickelt werden. Nachfolgend werden die zu erfüllenden strukturellen Voraussetzungen für die Synthese eines solchen Sensors nochmals zusammengefasst:
1. Verwendung eines Fluorophors, der eine hohe Signalintensität zeigt.
2. Analyt bzw. Erkennungseinheit soll sich im Abstand von wenigen Nanometern zum Signalgeber befinden, um bei jeder Detektionsreaktion die Signalintensität des Signalgebers beeinflussen zu können.
3. Die Detektionseinheit benötigt eine funktionelle Gruppe zur Immobilisierung. Immobilisierung kann z.B. durch Einbindung in ein Polymer erfolgen.
4. Der Fluorophor soll bei Änderung seiner lokalen Umgebung, durch Binden eines Löschers oder Änderung seiner Lösemittelumgebung seine Fluoreszenzeigenschaften drastisch ändern.
5. Die Reaktion sollte schnell und mit möglichst geringem apparativem Aufwand, am besten mit bloßem Auge zu verfolgen sein.
Für das ß-CD/Adamantan Modellsystem wurde ein Fluoreszenz Aus/An Sensor entwickelt, der bei Binden ß CD funktionalisierter Goldnanopartikel an das polymergebundene Adamantan die Fluoreszenz des RhodaminB Fluorophors effizient löscht und bei Verdrängung der Goldnanopartikel wieder zurück gewinnt. Dies konnte auch mit bloßem Auge verfolgt werden.
Für die Naphthalsäureimid Monomere, die mit NVCL copolymerisiert wurden, wurde abhängig von der lokalen Umgebung des Fluorophors eine unterschiedliche Verstärkung der Fluoreszenzintensität bei Überschreiten des Trübungspunktes des Polymers gefunden. Dabei zeigte sich, dass die Einführung eines Abstandshalters zwischen Polymerrückgrat und Fluorophor zu einer großen Fluoreszenz¬verstärkung führt, während sich ohne Abstandshalter die Fluoreszenzintensität bei Über¬schreiten des Trübungspunktes kaum ändert.
Two of the most controversial issues concerning the late Cenozoic evolution of the Andean orogen are the timing of uplift of the intraorogenic Puna plateau and its eastern border, the Eastern Cordillera, and ensuing changes in climatic and surface-process conditions in the intermontane basins of the NW-Argentine Andes. The Eastern Cordillera separates the internally drained, arid Puna from semi-arid intermontane basins and the humid sectors of the Andean broken foreland and the Subandean fold-and-thrust belt to the east. With elevations between 4,000 and 6,000 m the eastern flanks of the Andes form an efficient orographic barrier with westward-increasing elevation and asymmetric rainfall distribution and amount with respect to easterly moisture-bearing winds. This is mirrored by pronounced gradients in the efficiency of surface processes that erode and re-distribute sediment from the uplifting ranges. Although the overall pattern of deformation and uplift in this sector of the southern central Andes shows an eastward migration of deformation, a well-developed deformation front does not exist and uplift and associated erosion and sedimentary processes are highly disparate in space and time. In addition, periodic deformation within intermontane basins, and continued diachronous foreland uplifts associated with the reactivation of inherited basement structures furthermore make a rigorous assessment of the spatiotemporal uplift patterns difficult.
This thesis focuses on the tectonic evolution of the Eastern Cordillera of NW Argentina, the depositional history of its intermontane sedimentary basins, and the regional topographic evolution of the eastern flank of the Puna Plateau. The intermontane basins of the Eastern Cordillera and the adjacent morphotectonic provinces of the Sierras Pampeanas and the Santa Bárbara System are akin to reverse fault bounded, filled, and partly coalesced sedimentary basins of the Puna Plateau. In contrast to the Puna basins, however, which still form intact morphologic entities, repeated deformation, erosion, and re-filling have impacted the basins in the Eastern Cordillera. This has resulted in a rich stratigraphy of repeated basin fills, but many of these basins have retained vestiges of their early depositional history that may reach back in time when these areas were still part of a contiguous and undeformed foreland basin. Fortunately, these strata also contain abundant volcanic ashes that are not only important horizons to decipher tectono-sedimentary events through U-Pb geochronology and geochemical correlation, but they also represent terrestrial recorders of the hydrogen-isotope composition of ancient meteoric waters that can be compared to the isotopic composition of modern meteoric water. The ash horizons are thus unique recorders of past environmental conditions and lend themselves to tracking the development of rainfall barriers and tectonically forced climate and environmental change through time.
U-Pb zircon geochronology and paleocurrent reconstructions of conglomerate sequences in the Humahuaca Basin of the Eastern Cordillera at 23.5° S suggest that the basin was an integral part of a largely unrestricted depositional system until 4.2 Ma, which subsequently became progressively decoupled from the foreland by range uplifts to the east that forced easterly moisture-bearing winds to precipitate in increasingly eastward locations. Multiple cycles of severed hydrological conditions and drainage re-capture are identified together with these processes that were associated with basin filling and sediment evacuation, respectively. Moreover, systematic relationships among faults, regional unconformities and deformed landforms reveal a general pattern of intra-basin deformation that appears to be linked with basin-internal deformation during or subsequent to episodes of large-scale sediment removal. Some of these observations are supported by variations in the hydrogen stable isotope composition of volcanic glass from the Neogene to Quaternary sedimentary record, which can be related to spatiotemporal changes in topography and associated orographic effects. δDg values in the basin strata reveal two main trends associated with surface uplift in the catchment area between 6.0 and 3.5 Ma and the onset of semiarid conditions in the basin following the attainment of threshold elevations for effective orographic barriers to the east after 3.5 Ma. The disruption of sediment supply from western sources after 4.2 Ma and subsequent hinterland aridification, moreover, emphasize the possibility that these processes were related to lateral orogenic growth of the adjacent Puna Plateau. As a result of the hinterland aridification the regions in the orogen interior have been characterized by an inefficient fluvial system, which in turn has helped maintaining internal drainage conditions, sediment storage, and relief reduction within high-elevation basins.
The diachronous nature of basin formation and impacts on the fluvial system in the adjacent broken foreland is underscored by the results of detailed sediment provenance and paleocurrent analyses, as well as U-Pb zircon geochronology in the Lerma and Metán basins at ca. 25° S. This is particularly demonstrated by the isolated uplift of the Metán range at ~10 Ma, which is more than 50 km away from the presently active orogenic front along the eastern Puna margin and the Eastern Cordillera to the west. At about 5 Ma, Puna-sourced sediments disappear from the foreland record, documenting further range uplifts in the Eastern Cordillera and hydrological isolation of the neighboring Angastaco Basin from the foreland. Finally, during the late Pliocene and Quaternary, deformation has been accommodated across the entire foreland and is still active. To elucidate the interactions between tectonically controlled changes in elevation and their impact on atmospheric circulation processes in this region, this thesis provides additional, temporally well-constrained hydrogen stable isotope results of volcanic glass samples from the broken foreland, including the Angastaco Basin, and other intermontane basins farther south. The results suggest similar elevations of intermontane basins and the foreland sectors prior to ca. 7 Ma. In case of the Angastaco Basin the region was affected by km-scale surface uplift of the basin. A comparison with coeval isotope data collected from sedimentary sequences in the Puna plateau explains rapid shifts in the intermontane δDg record and supports the notion of recurring phases of enhanced deep convection during the Pliocene, and thus climatic conditions during the middle to late Pliocene similar to the present day.
Combined, field-based and isotope geochemical methods used in this study of the NW-Argentine Andes have thus helped to gain insight into the systematics, rate changes, interactions, and temporal characteristics among tectonically controlled deformation patterns, the build-up of topography impacting atmospheric processes, the distribution of rainfall, and resulting surface processes in a tectonically active mountain belt. Ultimately, this information is essential for a better understanding of the style and the rates at which non-collisional mountain belts evolve, including the development orogenic plateaus and their bordering flanks. The results presented in this study emphasize the importance of stable isotope records for paleoaltimetric and paleoenvironmental studies in mountain belts and furnishes important data for a rigorous interpretation of such records.
Parts without a whole?
(2015)
This explorative study gives a descriptive overview of what organizations do and experience when they say they practice design thinking. It looks at how the concept has been appropriated in organizations and also describes patterns of design thinking adoption. The authors use a mixed-method research design fed by two sources: questionnaire data and semi-structured personal expert interviews. The study proceeds in six parts: (1) design thinking¹s entry points into organizations; (2) understandings of the descriptor; (3) its fields of application and organizational localization; (4) its perceived impact; (5) reasons for its discontinuation or failure; and (6) attempts to measure its success. In conclusion the report challenges managers to be more conscious of their current design thinking practice. The authors suggest a co-evolution of the concept¹s introduction with innovation capability building and the respective changes in leadership approaches. It is argued that this might help in unfolding design thinking¹s hidden potentials as well as preventing unintended side-effects such as discontented teams or the dwindling authority of managers.
Introduction
We investigated blood glucose (BG) and hormone response to aerobic high-intensity interval exercise (HIIE) and moderate continuous exercise (CON) matched for mean load and duration in type 1 diabetes mellitus (T1DM).
Material and Methods
Seven trained male subjects with T1DM performed a maximal incremental exercise test and HIIE and CON at 3 different mean intensities below (A) and above (B) the first lactate turn point and below the second lactate turn point (C) on a cycle ergometer. Subjects were adjusted to ultra-long-acting insulin Degludec (Tresiba/ Novo Nordisk, Denmark). Before exercise, standardized meals were administered, and short-acting insulin dose was reduced by 25% (A), 50% (B), and 75% (C) dependent on mean exercise intensity. During exercise, BG, adrenaline, noradrenaline, dopamine, cortisol, glucagon, and insulin-like growth factor-1, blood lactate, heart rate, and gas exchange variables were measured. For 24 h after exercise, interstitial glucose was measured by continuous glucose monitoring system.
Results
BG decrease during HIIE was significantly smaller for B (p = 0.024) and tended to be smaller for A and C compared to CON. No differences were found for post-exercise interstitial glucose, acute hormone response, and carbohydrate utilization between HIIE and CON for A, B, and C. In HIIE, blood lactate for A (p = 0.006) and B (p = 0.004) and respiratory exchange ratio for A (p = 0.003) and B (p = 0.003) were significantly higher compared to CON but not for C.
Conclusion
Hypoglycemia did not occur during or after HIIE and CON when using ultra-long-acting insulin and applying our methodological approach for exercise prescription. HIIE led to a smaller BG decrease compared to CON, although both exercises modes were matched for mean load and duration, even despite markedly higher peak workloads applied in HIIE. Therefore, HIIE and CON could be safely performed in T1DM.
In living cells, there are always a plethora of processes taking place at the same time. Their precise regulation is the basis of cellular functions, since small failures can lead to severe dysfunctions. For a comprehensive understanding of intracellular homeostasis, simultaneous multiparameter detection is a versatile tool for revealing the spatial and temporal interactions of intracellular parameters. Here, a recently developed time-correlated single-photon counting (TCSPC) board was evaluated for simultaneous fluorescence and phosphorescence lifetime imaging microscopy (FLIM/PLIM). Therefore, the metabolic activity in insect salivary glands was investigated by recording ns-decaying intrinsic cellular fluorescence, mainly related to oxidized flavin adenine dinucleotide (FAD) and the μs-decaying phosphorescence of the oxygen-sensitive ruthenium-complex Kr341. Due to dopamine stimulation, the metabolic activity of salivary glands increased, causing a higher pericellular oxygen consumption and a resulting increase in Kr341 phosphorescence decay time. Furthermore, FAD fluorescence decay time decreased, presumably due to protein binding, thus inducing a quenching of FAD fluorescence decay time. Through application of the metabolic drugs antimycin and FCCP, the recorded signals could be assigned to a mitochondrial origin. The dopamine-induced changes could be observed in sequential FLIM and PLIM recordings, as well as in simultaneous FLIM/PLIM recordings using an intermediate TCSPC timing resolution.
Analphabetismus und Teilhabe
(2015)
Aus bildungstheoretisch-gesellschaftskritischer Perspektive stellt sich Lernen als soziales Handeln in gesellschaftlich-vermittelten Verhältnissen – Möglichkeiten wie auch Begrenzungen – dar. Funktionaler Analphabetismus ist mit einem bundesweiten Anteil von 14% der erwerbsfähigen Bevölkerung oder 7,5 Millionen Analphabeten in Deutschland nicht nur ein bildungspolitisches und -praktisches, sondern auch ein wissenschaftlich zu untersuchendes Phänomen. Es gibt zahlreiche Untersuchungen, die sich mit dieser Thematik auseinandersetzen und Anknüpfungspunkte für die vorliegende Studie bieten. Aus der Zielgruppenforschung beispielsweise ist bekannt, dass die Hauptadressaten der Männer, der Älteren und der Bildungsfernen nicht adäquat erreicht bzw. als Teilnehmende gewonnen werden. Aus der Teilnehmendenforschung sind Abbrüche und Drop-Outs bekannt.
Warum Analphabeten im Erwachsenenalter, also nach der Aneignung vielfältigster Bewältigungsstrategien, durch das sich das Phänomen einer direkten Sichtbarkeit entzieht, dennoch beginnen das Lesen und Schreiben (wieder) zu lernen, wird bislang weder bildungs- noch lerntheoretisch untersucht. Im Rahmen der vorliegenden Erwachsenenbildungsstudie werden genau diese Lernanlässe empirisch herausgearbeitet.
Als Heuristik wird auf eine subjekttheoretische Theoriefolie rekurriert, die sich in besonderer Weise eignet Lernbegründungen im Kontext gesellschaftlich verhafteter Biografien sichtbar zu machen. Lernforschung im Begründungsmodell muss dabei auf eine Methodik zurückgreifen, die die Perspektive des Subjekts, Bedeutungszusammenhänge und typische Sinnstrukturen hervorbringen kann. Daher wird ein auf Einzelfallstudien basierendes, qualitatives Forschungsdesign gewählt, das Daten aus der Erhebung mittels problemzentrierter Interviews bereitstellt, die eine Auswertung innerhalb der Forschungsstrategie der Grounded Theory erfahren und in einer empirisch begründeten Typenbildung münden. Dieses Design ermöglicht die Rekonstruktion typischer Lernanlässe und im Ergebnis die Entwicklung einer gegenstandsbezogene Theorie mittlerer Reichweite.
Aus der vorliegenden Bedeutungs-Begründungsanalyse konnten empirisch fünf Lernbegründungstypen ausdifferenziert werden, die sich im Spannungsverhältnis von Teilhabeausrichtung und Widersprüchlichkeit bewegen und in ihrer Komplexität mittels der drei Schlüsselkategorien Bedeutungsraum, Reflexion der sozialen Eingebundenheit und Kompetenzen sowie Lernen bzw. dem Erleben der Diskrepanzerfahrung zwischen Lesen-Wollen und Lesen-Können dargestellt werden. Das Spektrum der Lernbegründungstypen reicht von teilhabesicherndem resignierten Lernen, bei dem die Sicherung des bedrohten Status quo im Vordergrund steht und die Welt als nicht gestaltbar erlebt wird, bis hin zu vielschichtigem teilhabeerweiternden Lernen, das auf die Erweiterung der eigenen Handlungsmöglichkeiten zielt und die umfangreichste Reflexion der sozialen Eingebundenheit und Kompetenzen aufweist. Funktionale Analphabeten begründen ihr Lernen und Nicht-Lernen vor dem Hintergrund ihrer sozialen Situation, ihrer Begrenzungen und Möglichkeiten: Schriftsprachlernen erhält erst im Kontext gesellschaftlicher Teilhabe und dessen Reflexion eine Bedeutung.
Mit der Einordnung der Lernbegründungen funktionaler Analphabeten in: erstens, Diskurse der Bildungsbenachteiligung durch Exklusionsprozesse; zweitens, die lerntheoretische Bedeutung von Inklusionsprozessen und drittens, den internationalen Theorieansatz transformativen Lernens durch die Integration der Reflexionskategorie, erfolgt eine Erweiterung bildungs- und lerntheoretischer Ansätze. In dieser Arbeit werden Alphabetisierungs- und Erwachsenen-bildungsforschung verbunden und in den jeweiligen Diskurs integriert. Weitere Anschluss- und Verwertungsmöglichkeiten in der Bildungsforschung wären denkbar. Die Untersuchung von Lernbegründungen im Längsschnitt beispielsweise kann Transformationsprozesse rekonstruierbar machen und somit Erträge für eine Bildungsprozessforschung liefern. Bildungspraktisch können die Lernbegründungstypen einerseits der Teilnehmergewinnung dienen, andererseits Ausgangspunkt für reflexive Lernbegleitungskonzepte sein, die Lernbegründungen zur Sprache bringen und die soziale Eingebundenheit thematisieren und damit Lernprozesse unterstützen.
Jahresbericht 2014
(2015)
Das MenschenRechtsZentrum der Universität Potsdam (MRZ) beging im Jahr 2014 sein zwanzigjähriges Bestehen. Aus diesem Grund beschäftigt sich der aktuelle Jahresbericht nicht nur – wie ansonsten üblich – mit der spezifischen Organisationsstruktur und der Arbeit im Berichtszeitraum, sondern gibt einleitend einen knappen Überblick über die umfangreiche Tätigkeit des MRZ seit seiner Gründung. Diese Bilanz wird durch eine ausführliche Liste der Veranstaltungen und Schriftenreihen im Anhang vervollständigt.