TY - GEN A1 - Al Laban, Firas A1 - Reger, Martin A1 - Lucke, Ulrike T1 - Closing the Policy Gap in the Academic Bridge T2 - Zweitveröffentlichungen der Universität Potsdam : Mathematisch-Naturwissenschaftliche Reihe N2 - The highly structured nature of the educational sector demands effective policy mechanisms close to the needs of the field. That is why evidence-based policy making, endorsed by the European Commission under Erasmus+ Key Action 3, aims to make an alignment between the domains of policy and practice. Against this background, this article addresses two issues: First, that there is a vertical gap in the translation of higher-level policies to local strategies and regulations. Second, that there is a horizontal gap between educational domains regarding the policy awareness of individual players. This was analyzed in quantitative and qualitative studies with domain experts from the fields of virtual mobility and teacher training. From our findings, we argue that the combination of both gaps puts the academic bridge from secondary to tertiary education at risk, including the associated knowledge proficiency levels. We discuss the role of digitalization in the academic bridge by asking the question: which value does the involved stakeholders expect from educational policies? As a theoretical basis, we rely on the model of value co-creation for and by stakeholders. We describe the used instruments along with the obtained results and proposed benefits. Moreover, we reflect on the methodology applied, and we finally derive recommendations for future academic bridge policies. T3 - Zweitveröffentlichungen der Universität Potsdam : Mathematisch-Naturwissenschaftliche Reihe - 1310 KW - policy evaluation KW - higher education KW - virtual mobility KW - teacher training Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:kobv:517-opus4-583572 SN - 1866-8372 IS - 1310 ER - TY - THES A1 - Cichalla, Anika Katleen T1 - Ein konstruktivistisches Modell für die Didaktik der Informatik im Bachelorstudium T1 - A constructivistic model for the didactics of computational science in bachelor studies N2 - Lehrende in der Lehrkräfteausbildung sind stets damit konfrontiert, dass sie den Studierenden innovative Methoden modernen Schulunterrichts traditionell rezipierend vorstellen. In Deutschland gibt es circa 40 Universitäten, die Informatik mit Lehramtsbezug ausbilden. Allerdings gibt es nur wenige Konzepte, die sich mit der Verbindung von Bildungswissenschaften und der Informatik mit ihrer Didaktik beschäftigen und keine Konzepte, die eine konstruktivistische Lehre in der Informatik verfolgen. Daher zielt diese Masterarbeit darauf ab, diese Lücke aufgreifen und anhand des „Didaktik der Informatik I“ Moduls der Universität Potsdam ein Modell zur konstruktivistischen Hochschullehre zu entwickeln. Dabei soll ein bestehendes konstruktivistisches Lehrmodell auf die Informatikdidaktik übertragen und Elemente zur Verbindung von Bildungswissenschaften, Fachwissenschaften und Fachdidaktiken mit einbezogen werden. Dies kann eine Grundlage für die Planung von Informatikdidaktischen Modulen bieten, aber auch als Inspiration zur Übertragung bestehender innovativer Lehrkonzepte auf andere Fachdidaktiken dienen. Um ein solches konstruktivistisches Lehr-Lern-Modell zu erstellen, wird zunächst der Zusammenhang von Bildungswissenschaften, Fachwissenschaften und Fachdidaktiken erläutert und anschließend die Notwendigkeit einer Vernetzung hervorgehoben. Hieran folgt eine Darstellung zu relevanten Lerntheorien und bereits entwickelten innovativen Lernkonzepten. Anknüpfend wird darauf eingegangen, welche Anforderungen die Kultusminister- Konferenz an die Ausbildung von Lehrkräften stellt und wie diese Ausbildung für die Informatik momentan an der Universität Potsdam erfolgt. Aus allen Erkenntnissen heraus werden Anforderungen an ein konstruktivistisches Lehrmodell festgelegt. Unter Berücksichtigung der Voraussetzungen der Studienordnung für das Lehramt Informatik wird anschließend ein Modell für konstruktivistische Informatikdidaktik vorgestellt. Weiterführende Forschung könnte sich damit auseinandersetzen, inwiefern sich die Motivation und Leistung im vergleich zum ursprünglichen Modul ändert und ob die Kompetenzen zur Unterrichtsplanung und Unterrichtsgestaltung durch das neue Modulkonzept stärker ausgebaut werden können. N2 - Teachers in teacher training are always confronted with the fact that they present innovative methods of modern school teaching to students in a traditionally receptive way. In Germany, there are about 40 universities that train computational science with a focus on teaching. However, there are only a few concepts that deal with the connection of educational science and computer science with its didactics and no concepts that pursue constructivist teaching in computational science. Therefore, this master thesis aims to address this gap and to develop a model for constructivist university teaching based on the "Didactics of Computational Science I" module at the University of Potsdam. An existing constructivist teaching model is to be transferred to computational science didactics and elements for the connection of general pedagogy, scientific theory and didactics are to be included. This can provide a basis for planning computational science didactic modules, but also serve as inspiration for transferring existing innovative teaching concepts to other subject didactics. In order to create such a constructivist teaching-learning model, the interrelationship of general pedagogy, scientific theory and didactics is first explained and then the necessity of networking is emphasized. This is followed by a presentation of relevant learning theories and innovative learning concepts already developed. Subsequently, the requirements of the Standing Conference of the Ministers of Education and Cultural Affairs (Kultusministerkonferenz) for the training of teachers and how this training for computer science is currently carried out at the University of Potsdam are discussed. From all findings, requirements for a constructivist teaching model are defined. Taking into account the requirements of the study regulations for the computer science teaching profession, a model for constructivist computer science didactics is then presented. Further research could address the extent to which motivation and performance change in comparison to the original module and whether the competencies for lesson planning and lesson design can be more developed on base of the new module concept. KW - education KW - university education KW - teacher training KW - Hochschulbildung KW - Lehrkräfteausbildung KW - Konstruktivismus KW - construktivism KW - Informatik KW - Informatikdidaktik KW - Computational Science KW - didactics Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:kobv:517-opus4-550710 ER - TY - JOUR A1 - Lorenz, Claas A1 - Clemens, Vera Elisabeth A1 - Schrötter, Max A1 - Schnor, Bettina T1 - Continuous verification of network security compliance JF - IEEE transactions on network and service management N2 - Continuous verification of network security compliance is an accepted need. Especially, the analysis of stateful packet filters plays a central role for network security in practice. But the few existing tools which support the analysis of stateful packet filters are based on general applicable formal methods like Satifiability Modulo Theories (SMT) or theorem prover and show runtimes in the order of minutes to hours making them unsuitable for continuous compliance verification. In this work, we address these challenges and present the concept of state shell interweaving to transform a stateful firewall rule set into a stateless rule set. This allows us to reuse any fast domain specific engine from the field of data plane verification tools leveraging smart, very fast, and domain specialized data structures and algorithms including Header Space Analysis (HSA). First, we introduce the formal language FPL that enables a high-level human-understandable specification of the desired state of network security. Second, we demonstrate the instantiation of a compliance process using a verification framework that analyzes the configuration of complex networks and devices - including stateful firewalls - for compliance with FPL policies. Our evaluation results show the scalability of the presented approach for the well known Internet2 and Stanford benchmarks as well as for large firewall rule sets where it outscales state-of-the-art tools by a factor of over 41. KW - Security KW - Tools KW - Network security KW - Engines KW - Benchmark testing; KW - Analytical models KW - Scalability KW - Network KW - security KW - compliance KW - formal KW - verification Y1 - 2021 U6 - https://doi.org/10.1109/TNSM.2021.3130290 SN - 1932-4537 VL - 19 IS - 2 SP - 1729 EP - 1745 PB - Institute of Electrical and Electronics Engineers CY - New York ER - TY - JOUR A1 - Prasse, Paul A1 - Iversen, Pascal A1 - Lienhard, Matthias A1 - Thedinga, Kristina A1 - Bauer, Christopher A1 - Herwig, Ralf A1 - Scheffer, Tobias T1 - Matching anticancer compounds and tumor cell lines by neural networks with ranking loss JF - NAR: genomics and bioinformatics N2 - Computational drug sensitivity models have the potential to improve therapeutic outcomes by identifying targeted drug components that are likely to achieve the highest efficacy for a cancer cell line at hand at a therapeutic dose. State of the art drug sensitivity models use regression techniques to predict the inhibitory concentration of a drug for a tumor cell line. This regression objective is not directly aligned with either of these principal goals of drug sensitivity models: We argue that drug sensitivity modeling should be seen as a ranking problem with an optimization criterion that quantifies a drug's inhibitory capacity for the cancer cell line at hand relative to its toxicity for healthy cells. We derive an extension to the well-established drug sensitivity regression model PaccMann that employs a ranking loss and focuses on the ratio of inhibitory concentration and therapeutic dosage range. We find that the ranking extension significantly enhances the model's capability to identify the most effective anticancer drugs for unseen tumor cell profiles based in on in-vitro data. Y1 - 2022 U6 - https://doi.org/10.1093/nargab/lqab128 SN - 2631-9268 VL - 4 IS - 1 PB - Oxford Univ. Press CY - Oxford ER - TY - JOUR A1 - Steinert, Fritjof A1 - Stabernack, Benno T1 - Architecture of a low latency H.264/AVC video codec for robust ML based image classification how region of interests can minimize the impact of coding artifacts JF - Journal of Signal Processing Systems for Signal, Image, and Video Technology N2 - The use of neural networks is considered as the state of the art in the field of image classification. A large number of different networks are available for this purpose, which, appropriately trained, permit a high level of classification accuracy. Typically, these networks are applied to uncompressed image data, since a corresponding training was also carried out using image data of similar high quality. However, if image data contains image errors, the classification accuracy deteriorates drastically. This applies in particular to coding artifacts which occur due to image and video compression. Typical application scenarios for video compression are narrowband transmission channels for which video coding is required but a subsequent classification is to be carried out on the receiver side. In this paper we present a special H.264/Advanced Video Codec (AVC) based video codec that allows certain regions of a picture to be coded with near constant picture quality in order to allow a reliable classification using neural networks, whereas the remaining image will be coded using constant bit rate. We have combined this feature with the ability to run with lowest latency properties, which is usually also required in remote control applications scenarios. The codec has been implemented as a fully hardwired High Definition video capable hardware architecture which is suitable for Field Programmable Gate Arrays. KW - H.264 KW - Advanced Video Codec (AVC) KW - Low Latency KW - Region of Interest KW - Machine Learning KW - Inference KW - FPGA KW - Hardware accelerator Y1 - 2022 U6 - https://doi.org/10.1007/s11265-021-01727-2 SN - 1939-8018 SN - 1939-8115 VL - 94 IS - 7 SP - 693 EP - 708 PB - Springer CY - New York ER - TY - JOUR A1 - Marco Figuera, Ramiro A1 - Riedel, Christian A1 - Rossi, Angelo Pio A1 - Unnithan, Vikram T1 - Depth to diameter analysis on small simple craters at the lunar south pole - possible implications for ice harboring JF - Remote sensing N2 - In this paper, we present a study comparing the depth to diameter (d/D) ratio of small simple craters (200-1000 m) of an area between -88.5 degrees to -90 degrees latitude at the lunar south pole containing Permanent Shadowed Regions (PSRs) versus craters without PSRs. As PSRs can reach temperatures of 110 K and are capable of harboring volatiles, especially water ice, we analyzed the relationship of depth versus diameter ratios and its possible implications for harboring water ice. Variations in the d/D ratios can also be caused by other processes such as degradation, isostatic adjustment, or differences in surface properties. The conducted d/D ratio analysis suggests that a differentiation between craters containing PSRs versus craters without PSRs occurs. Thus, a possible direct relation between d/D ratio, PSRs, and water ice harboring might exist. Our results suggest that differences in the target's surface properties may explain the obtained results. The resulting d/D ratios of craters with PSRs can help to select target areas for future In-Situ Resource Utilization (ISRU) missions. KW - craters KW - lunar exploration KW - ice harboring Y1 - 2022 U6 - https://doi.org/10.3390/rs14030450 SN - 2072-4292 VL - 14 IS - 3 PB - MDPI CY - Basel ER - TY - JOUR A1 - Abdelwahab, Ahmed A1 - Landwehr, Niels T1 - Deep Distributional Sequence Embeddings Based on a Wasserstein Loss JF - Neural processing letters N2 - Deep metric learning employs deep neural networks to embed instances into a metric space such that distances between instances of the same class are small and distances between instances from different classes are large. In most existing deep metric learning techniques, the embedding of an instance is given by a feature vector produced by a deep neural network and Euclidean distance or cosine similarity defines distances between these vectors. This paper studies deep distributional embeddings of sequences, where the embedding of a sequence is given by the distribution of learned deep features across the sequence. The motivation for this is to better capture statistical information about the distribution of patterns within the sequence in the embedding. When embeddings are distributions rather than vectors, measuring distances between embeddings involves comparing their respective distributions. The paper therefore proposes a distance metric based on Wasserstein distances between the distributions and a corresponding loss function for metric learning, which leads to a novel end-to-end trainable embedding model. We empirically observe that distributional embeddings outperform standard vector embeddings and that training with the proposed Wasserstein metric outperforms training with other distance functions. KW - Metric learning KW - Sequence embeddings KW - Deep learning Y1 - 2022 U6 - https://doi.org/10.1007/s11063-022-10784-y SN - 1370-4621 SN - 1573-773X PB - Springer CY - Dordrecht ER - TY - JOUR A1 - Tran, Son Cao A1 - Pontelli, Enrico A1 - Balduccini, Marcello A1 - Schaub, Torsten T1 - Answer set planning BT - a survey JF - Theory and practice of logic programming N2 - Answer Set Planning refers to the use of Answer Set Programming (ASP) to compute plans, that is, solutions to planning problems, that transform a given state of the world to another state. The development of efficient and scalable answer set solvers has provided a significant boost to the development of ASP-based planning systems. This paper surveys the progress made during the last two and a half decades in the area of answer set planning, from its foundations to its use in challenging planning domains. The survey explores the advantages and disadvantages of answer set planning. It also discusses typical applications of answer set planning and presents a set of challenges for future research. KW - planning KW - knowledge representation and reasoning KW - logic programming Y1 - 2022 U6 - https://doi.org/10.1017/S1471068422000072 SN - 1471-0684 SN - 1475-3081 PB - Cambridge University Press CY - New York ER - TY - JOUR A1 - Michallek, Florian A1 - Genske, Ulrich A1 - Niehues, Stefan Markus A1 - Hamm, Bernd A1 - Jahnke, Paul T1 - Deep learning reconstruction improves radiomics feature stability and discriminative power in abdominal CT imaging BT - a phantom study JF - European Radiology N2 - Objectives To compare image quality of deep learning reconstruction (AiCE) for radiomics feature extraction with filtered back projection (FBP), hybrid iterative reconstruction (AIDR 3D), and model-based iterative reconstruction (FIRST). Methods Effects of image reconstruction on radiomics features were investigated using a phantom that realistically mimicked a 65-year-old patient's abdomen with hepatic metastases. The phantom was scanned at 18 doses from 0.2 to 4 mGy, with 20 repeated scans per dose. Images were reconstructed with FBP, AIDR 3D, FIRST, and AiCE. Ninety-three radiomics features were extracted from 24 regions of interest, which were evenly distributed across three tissue classes: normal liver, metastatic core, and metastatic rim. Features were analyzed in terms of their consistent characterization of tissues within the same image (intraclass correlation coefficient >= 0.75), discriminative power (Kruskal-Wallis test p value < 0.05), and repeatability (overall concordance correlation coefficient >= 0.75). Results The median fraction of consistent features across all doses was 6%, 8%, 6%, and 22% with FBP, AIDR 3D, FIRST, and AiCE, respectively. Adequate discriminative power was achieved by 48%, 82%, 84%, and 92% of features, and 52%, 20%, 17%, and 39% of features were repeatable, respectively. Only 5% of features combined consistency, discriminative power, and repeatability with FBP, AIDR 3D, and FIRST versus 13% with AiCE at doses above 1 mGy and 17% at doses >= 3 mGy. AiCE was the only reconstruction technique that enabled extraction of higher-order features. Conclusions AiCE more than doubled the yield of radiomics features at doses typically used clinically. Inconsistent tissue characterization within CT images contributes significantly to the poor stability of radiomics features. KW - Tomography KW - X-ray computed KW - Phantoms KW - imaging KW - Liver neoplasms KW - Algorithms KW - Reproducibility of results Y1 - 2022 U6 - https://doi.org/10.1007/s00330-022-08592-y SN - 1432-1084 VL - 32 IS - 7 SP - 4587 EP - 4595 PB - Springer CY - New York ER - TY - JOUR A1 - Bandyopadhyay, Soumyadip A1 - Sarkar, Dipankar A1 - Mandal, Chittaranjan A1 - Giese, Holger T1 - Translation validation of coloured Petri net models of programs on integers JF - Acta informatica N2 - Programs are often subjected to significant optimizing and parallelizing transformations based on extensive dependence analysis. Formal validation of such transformations needs modelling paradigms which can capture both control and data dependences in the program vividly. Being value-based with an inherent scope of capturing parallelism, the untimed coloured Petri net (CPN) models, reported in the literature, fit the bill well; accordingly, they are likely to be more convenient as the intermediate representations (IRs) of both the source and the transformed codes for translation validation than strictly sequential variable-based IRs like sequential control flow graphs (CFGs). In this work, an efficient path-based equivalence checking method for CPN models of programs on integers is presented. Extensive experimentation has been carried out on several sequential and parallel examples. Complexity and correctness issues have been treated rigorously for the method. Y1 - 2022 U6 - https://doi.org/10.1007/s00236-022-00419-z SN - 0001-5903 SN - 1432-0525 VL - 59 IS - 6 SP - 725 EP - 759 PB - Springer CY - New York ER - TY - JOUR A1 - Chen, Junchao A1 - Lange, Thomas A1 - Andjelkovic, Marko A1 - Simevski, Aleksandar A1 - Lu, Li A1 - Krstić, Miloš T1 - Solar particle event and single event upset prediction from SRAM-based monitor and supervised machine learning JF - IEEE transactions on emerging topics in computing / IEEE Computer Society, Institute of Electrical and Electronics Engineers N2 - The intensity of cosmic radiation may differ over five orders of magnitude within a few hours or days during the Solar Particle Events (SPEs), thus increasing for several orders of magnitude the probability of Single Event Upsets (SEUs) in space-borne electronic systems. Therefore, it is vital to enable the early detection of the SEU rate changes in order to ensure timely activation of dynamic radiation hardening measures. In this paper, an embedded approach for the prediction of SPEs and SRAM SEU rate is presented. The proposed solution combines the real-time SRAM-based SEU monitor, the offline-trained machine learning model and online learning algorithm for the prediction. With respect to the state-of-the-art, our solution brings the following benefits: (1) Use of existing on-chip data storage SRAM as a particle detector, thus minimizing the hardware and power overhead, (2) Prediction of SRAM SEU rate one hour in advance, with the fine-grained hourly tracking of SEU variations during SPEs as well as under normal conditions, (3) Online optimization of the prediction model for enhancing the prediction accuracy during run-time, (4) Negligible cost of hardware accelerator design for the implementation of selected machine learning model and online learning algorithm. The proposed design is intended for a highly dependable and self-adaptive multiprocessing system employed in space applications, allowing to trigger the radiation mitigation mechanisms before the onset of high radiation levels. KW - Machine learning KW - Single event upsets KW - Random access memory KW - monitoring KW - machine learning algorithms KW - predictive models KW - space missions KW - solar particle event KW - single event upset KW - machine learning KW - online learning KW - hardware accelerator KW - reliability KW - self-adaptive multiprocessing system Y1 - 2022 U6 - https://doi.org/10.1109/TETC.2022.3147376 SN - 2168-6750 VL - 10 IS - 2 SP - 564 EP - 580 PB - Institute of Electrical and Electronics Engineers CY - [New York, NY] ER - TY - JOUR A1 - Breitenreiter, Anselm A1 - Andjelković, Marko A1 - Schrape, Oliver A1 - Krstić, Miloš T1 - Fast error propagation probability estimates by answer set programming and approximate model counting JF - IEEE Access N2 - We present a method employing Answer Set Programming in combination with Approximate Model Counting for fast and accurate calculation of error propagation probabilities in digital circuits. By an efficient problem encoding, we achieve an input data format similar to a Verilog netlist so that extensive preprocessing is avoided. By a tight interconnection of our application with the underlying solver, we avoid iterating over fault sites and reduce calls to the solver. Several circuits were analyzed with varying numbers of considered cycles and different degrees of approximation. Our experiments show, that the runtime can be reduced by approximation by a factor of 91, whereas the error compared to the exact result is below 1%. KW - Circuit faults KW - Integrated circuit modeling KW - Programming KW - Analytical models KW - Search problems KW - Flip-flops KW - Encoding KW - Answer set programming KW - approximate model counting KW - error propagation KW - radhard design KW - reliability analysis KW - selective fault tolerance KW - single event upsets Y1 - 2022 U6 - https://doi.org/10.1109/ACCESS.2022.3174564 SN - 2169-3536 VL - 10 SP - 51814 EP - 51825 PB - Inst. of Electr. and Electronics Engineers CY - Piscataway ER - TY - JOUR A1 - Andjelkovic, Marko A1 - Simevski, Aleksandar A1 - Chen, Junchao A1 - Schrape, Oliver A1 - Stamenkovic, Zoran A1 - Krstić, Miloš A1 - Ilic, Stefan A1 - Ristic, Goran A1 - Jaksic, Aleksandar A1 - Vasovic, Nikola A1 - Duane, Russell A1 - Palma, Alberto J. A1 - Lallena, Antonio M. A1 - Carvajal, Miguel A. T1 - A design concept for radiation hardened RADFET readout system for space applications JF - Microprocessors and microsystems N2 - Instruments for measuring the absorbed dose and dose rate under radiation exposure, known as radiation dosimeters, are indispensable in space missions. They are composed of radiation sensors that generate current or voltage response when exposed to ionizing radiation, and processing electronics for computing the absorbed dose and dose rate. Among a wide range of existing radiation sensors, the Radiation Sensitive Field Effect Transistors (RADFETs) have unique advantages for absorbed dose measurement, and a proven record of successful exploitation in space missions. It has been shown that the RADFETs may be also used for the dose rate monitoring. In that regard, we propose a unique design concept that supports the simultaneous operation of a single RADFET as absorbed dose and dose rate monitor. This enables to reduce the cost of implementation, since the need for other types of radiation sensors can be minimized or eliminated. For processing the RADFET's response we propose a readout system composed of analog signal conditioner (ASC) and a self-adaptive multiprocessing system-on-chip (MPSoC). The soft error rate of MPSoC is monitored in real time with embedded sensors, allowing the autonomous switching between three operating modes (high-performance, de-stress and fault-tolerant), according to the application requirements and radiation conditions. KW - RADFET KW - Radiation hardness KW - Absorbed dose KW - Dose rate KW - Self-adaptive MPSoC Y1 - 2022 U6 - https://doi.org/10.1016/j.micpro.2022.104486 SN - 0141-9331 SN - 1872-9436 VL - 90 PB - Elsevier CY - Amsterdam ER - TY - JOUR A1 - Ristic, Goran S. A1 - Ilic, Stefan D. A1 - Andjelkovic, Marko S. A1 - Duane, Russell A1 - Palma, Alberto J. A1 - Lalena, Antonio M. A1 - Krstić, Miloš A1 - Jaksic, Aleksandar B. T1 - Sensitivity and fading of irradiated RADFETs with different gate voltages JF - Nuclear Instruments and Methods in Physics Research Section A N2 - The radiation-sensitive field-effect transistors (RADFETs) with an oxide thickness of 400 nm are irradiated with gate voltages of 2, 4 and 6 V, and without gate voltage. A detailed analysis of the mechanisms responsible for the creation of traps during irradiation is performed. The creation of the traps in the oxide, near and at the silicon/silicon-dioxide (Si/SiO2) interface during irradiation is modelled very well. This modelling can also be used for other MOS transistors containing SiO2. The behaviour of radiation traps during postirradiation annealing is analysed, and the corresponding functions for their modelling are obtained. The switching traps (STs) do not have significant influence on threshold voltage shift, and two radiation-induced trap types fit the fixed traps (FTs) very well. The fading does not depend on the positive gate voltage applied during irradiation, but it is twice lower in case there is no gate voltage. A new dosimetric parameter, called the Golden Ratio (GR), is proposed, which represents the ratio between the threshold voltage shift after irradiation and fading after spontaneous annealing. This parameter can be useful for comparing MOS dosimeters. KW - pMOS radiation dosimeter KW - RADFETs KW - irradiation KW - sensitivity KW - annealing KW - fading Y1 - 2022 U6 - https://doi.org/10.1016/j.nima.2022.166473 SN - 0168-9002 SN - 1872-9576 VL - 1029 PB - Elsevier CY - Amsterdam ER - TY - THES A1 - Klockmann, Alexander T1 - Modifizierte Unidirektionale Codes für Speicherfehler N2 - Das Promotionsvorhaben verfolgt das Ziel, die Zuverlässigkeit der Datenspeicherung und die Speicherdichte von neu entwickelten Speichern (Emerging Memories) mit Multi-Level-Speicherzellen zu verbessern bzw. zu erhöhen. Hierfür werden Codes zur Erkennung von unidirektionalen Fehlern analysiert, modifiziert und neu entwickelt, um sie innerhalb der neuen Speicher anwenden zu können. Der Fokus liegt dabei auf sog. Berger-Codes und m-aus-n-Codes. Da Multi-Level-Speicherzellen nicht mehr binär, sondern mit mehreren Leveln arbeiten, können bisher verwendete Codes nicht mehr verwendet werden, bzw. müssen entsprechend angepasst werden. Auf Basis der Berger-Codes und m-aus-n-Codes werden in dieser Arbeit neue Codes abgeleitet, welche in der Lage sind, Daten auch in mehrwertigen Systemen zu schützen. KW - Fehlererkennung KW - Codierungstheorie KW - Speicher KW - unidirektionale Fehler Y1 - 2022 ER - TY - THES A1 - Repp, Leo T1 - Extending the automatic theorem prover nanoCoP with arithmetic procedures T1 - Erweiterung des automatischen Theorembeweisers nanoCoP um Arithmetik und Gleichheit behandelnde Verfahren N2 - In dieser Bachelorarbeit implementiere ich den automatischen Theorembeweiser nanoCoP-Ω. Es handelt sich bei diesem neuen System um das Ergebnis einer Portierung von Arithmetik-behandelnden Prozeduren aus dem automatischen Theorembeweiser mit Arithmetik leanCoP-Ω in das System nanoCoP 2.0. Dazu wird zuerst der mathematische Hintergrund zu automatischen Theorembeweisern und Arithmetik gegeben. Ich stelle die Vorgängerprojekte leanCoP, nanoCoP und leanCoP-Ω vor, auf dessen Vorlage nanoCoP-Ω entwickelt wurde. Es folgt eine ausführliche Erklärung der Konzepte, um welche der nicht-klausale Konnektionskalkül erweitert werden muss, um eine Behandlung von arithmetischen Ausdrücken und Gleichheiten in den Kalkül zu integrieren, sowie eine Beschreibung der Implementierung dieser Konzepte in nanoCoP-Ω. Als letztes folgt eine experimentelle Evaluation von nanoCoP-Ω. Es wurde ein ausführlicher Vergleich von Laufzeit und Anzahl gelöster Probleme im Vergleich zum ähnlich aufgebauten Theorembeweiser leanCoP-Ω auf Basis der TPTP-Benchmark durchgeführt. Ich komme zu dem Ergebnis, dass nanoCoP-Ω deutlich schneller ist als leanCoP-Ω ist, jedoch weniger gut geeignet für größere Probleme. Zudem konnte ich feststellen, dass nanoCoP-Ω falsche Beweise liefern kann. Ich bespreche, wie dieses Problem gelöst werden kann, sowie einige mögliche Optimierungen und Erweiterungen des Beweissystems. N2 - In this bachelor’s thesis I implement the automatic theorem prover nanoCoP-Ω. This system is the result of porting arithmetic and equality handling procedures first introduced in the automatic theorem prover with arithmetic leanCoP-Ω into the similar system nanoCoP 2.0. To understand these procedures, I first introduce the mathematical background to both automatic theorem proving and arithmetic expressions. I present the predecessor projects leanCoP, nanoCoP and leanCoP-Ω, out of which nanCoP-Ω was developed. This is followed by an extensive description of the concepts the non-clausal connection calculus needed to be extended by, to allow for proving arithmetic expressions and equalities, as well as of their implementation into nanoCoP-Ω. An extensive comparison between both the runtimes and the number of solved problems of the systems nanoCoP-Ω and leanCoP-Ω was made. I come to the conclusion, that nanoCoP-Ω is considerably faster than leanCoP-Ω for small problems, though less well suited for larger problems. Additionally, I was able to construct a non-theorem that nanoCoP-Ω generates a false proof for. I discuss how this pressing issue could be resolved, as well as some possible optimizations and expansions of the system. KW - automatic theorem prover KW - leanCoP KW - connection calculus KW - tptp KW - arithmetic procedures KW - equality KW - omega KW - arithmethische Prozeduren KW - automatisierter Theorembeweiser KW - Konnektionskalkül KW - Gleichheit KW - leanCoP KW - Omega KW - TPTP Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:kobv:517-opus4-576195 ER - TY - THES A1 - Kaminski, Roland T1 - Complex reasoning with answer set programming N2 - Answer Set Programming (ASP) allows us to address knowledge-intensive search and optimization problems in a declarative way due to its integrated modeling, grounding, and solving workflow. A problem is modeled using a rule based language and then grounded and solved. Solving results in a set of stable models that correspond to solutions of the modeled problem. In this thesis, we present the design and implementation of the clingo system---perhaps, the most widely used ASP system. It features a rich modeling language originating from the field of knowledge representation and reasoning, efficient grounding algorithms based on database evaluation techniques, and high performance solving algorithms based on Boolean satisfiability (SAT) solving technology. The contributions of this thesis lie in the design of the modeling language, the design and implementation of the grounding algorithms, and the design and implementation of an Application Programmable Interface (API) facilitating the use of ASP in real world applications and the implementation of complex forms of reasoning beyond the traditional ASP workflow. KW - Answer Set Programming KW - Declarative Problem Solving KW - Grounding Theory KW - Preference Handling KW - Answer Set Solving modulo Theories KW - Temporal Answer Set Solving Y1 - 2023 ER - TY - CHAP A1 - Desel, Jörg A1 - Opel, Simone A1 - Siegeris, Juliane A1 - Draude, Claude A1 - Weber, Gerhard A1 - Schell, Timon A1 - Schwill, Andreas A1 - Thorbrügge, Carsten A1 - Schäfer, Len Ole A1 - Netzer, Cajus Marian A1 - Gerstenberger, Dietrich A1 - Winkelnkemper, Felix A1 - Schulte, Carsten A1 - Böttcher, Axel A1 - Thurner, Veronika A1 - Häfner, Tanja A1 - Ottinger, Sarah A1 - Große-Bölting, Gregor A1 - Scheppach, Lukas A1 - Mühling, Andreas A1 - Baberowski, David A1 - Leonhardt, Thiemo A1 - Rentsch, Susanne A1 - Bergner, Nadine A1 - Bonorden, Leif A1 - Stemme, Jonas A1 - Hoppe, Uwe A1 - Weicker, Karsten A1 - Bender, Esther A1 - Barbas, Helena A1 - Hamann, Fabian A1 - Soll, Marcus A1 - Sitzmann, Daniel ED - Desel, Jörg ED - Opel, Simone ED - Siegeris, Juliane T1 - Hochschuldidaktik Informatik HDI 2021 BT - 9. Fachtagung des GI-Fachbereichs Informatik und Ausbildung/Didaktik der Informatik 15.–16. September 2021 in Dortmund T2 - Commentarii informaticae didacticae N2 - Die Fachtagungen HDI (Hochschuldidaktik Informatik) beschäftigen sich mit den unterschiedlichen Aspekten informatischer Bildung im Hochschulbereich. Neben den allgemeinen Themen wie verschiedenen Lehr- und Lernformen, dem Einsatz von Informatiksystemen in der Hochschullehre oder Fragen der Gewinnung von geeigneten Studierenden, deren Kompetenzerwerb oder auch der Betreuung der Studierenden widmet sich die HDI immer auch einem Schwerpunktthema. Im Jahr 2021 war dies die Berücksichtigung von Diversität in der Lehre. Diskutiert wurden beispielsweise die Einbeziehung von besonderen fachlichen und überfachlichen Kompetenzen Studierender, der Unterstützung von Durchlässigkeit aus nichtakademischen Berufen, aber auch die Gestaltung inklusiver Lehr- und Lernszenarios, Aspekte des Lebenslangen Lernens oder sich an die Diversität von Studierenden adaptierte oder adaptierende Lehrsysteme. Dieser Band enthält ausgewählte Beiträge der 9. Fachtagung 2021, die in besonderer Weise die Konferenz und die dort diskutierten Themen repräsentieren. T3 - Commentarii informaticae didacticae (CID) - 13 KW - Hochschuldidaktik KW - Informatikdidaktik KW - HDI KW - Hochschullehre KW - digitale Hochschullehre KW - Diversität KW - Heterogenität KW - Lebenslanges Lernen KW - Informatikstudium KW - Didaktische Konzepte KW - Assessment Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:kobv:517-opus4-565070 SN - 978-3-86956-548-4 SN - 1868-0844 SN - 2191-1940 IS - 13 PB - Universitätsverlag Potsdam CY - Potsdam ER - TY - THES A1 - Middelanis, Robin T1 - Global response to local extremes—a storyline approach on economic loss propagation from weather extremes T1 - Globale Reaktion auf lokale Extreme — ein Storyline-Ansatz zu ökonomischer Schadensausbreitung aufgrund von Wetterextremen N2 - Due to anthropogenic greenhouse gas emissions, Earth’s average surface temperature is steadily increasing. As a consequence, many weather extremes are likely to become more frequent and intense. This poses a threat to natural and human systems, with local impacts capable of destroying exposed assets and infrastructure, and disrupting economic and societal activity. Yet, these effects are not locally confined to the directly affected regions, as they can trigger indirect economic repercussions through loss propagation along supply chains. As a result, local extremes yield a potentially global economic response. To build economic resilience and design effective adaptation measures that mitigate adverse socio-economic impacts of ongoing climate change, it is crucial to gain a comprehensive understanding of indirect impacts and the underlying economic mechanisms. Presenting six articles in this thesis, I contribute towards this understanding. To this end, I expand on local impacts under current and future climate, the resulting global economic response, as well as the methods and tools to analyze this response. Starting with a traditional assessment of weather extremes under climate change, the first article investigates extreme snowfall in the Northern Hemisphere until the end of the century. Analyzing an ensemble of global climate model projections reveals an increase of the most extreme snowfall, while mean snowfall decreases. Assessing repercussions beyond local impacts, I employ numerical simulations to compute indirect economic effects from weather extremes with the numerical agent-based shock propagation model Acclimate. This model is used in conjunction with the recently emerged storyline framework, which involves analyzing the impacts of a particular reference extreme event and comparing them to impacts in plausible counterfactual scenarios under various climate or socio-economic conditions. Using this approach, I introduce three primary storylines that shed light on the complex mechanisms underlying economic loss propagation. In the second and third articles of this thesis, I analyze storylines for the historical Hurricanes Sandy (2012) and Harvey (2017) in the USA. For this, I first estimate local economic output losses and then simulate the resulting global economic response with Acclimate. The storyline for Hurricane Sandy thereby focuses on global consumption price anomalies and the resulting changes in consumption. I find that the local economic disruption leads to a global wave-like economic price ripple, with upstream effects propagating in the supplier direction and downstream effects in the buyer direction. Initially, an upstream demand reduction causes consumption price decreases, followed by a downstream supply shortage and increasing prices, before the anomalies decay in a normalization phase. A dominant upstream or downstream effect leads to net consumption gains or losses of a region, respectively. Moreover, I demonstrate that a longer direct economic shock intensifies the downstream effect for many regions, leading to an overall consumption loss. The third article of my thesis builds upon the developed loss estimation method by incorporating projections to future global warming levels. I use these projections to explore how the global production response to Hurricane Harvey would change under further increased global warming. The results show that, while the USA is able to nationally offset direct losses in the reference configuration, other countries have to compensate for increasing shares of counterfactual future losses. This compensation is mainly achieved by large exporting countries, but gradually shifts towards smaller regions. These findings not only highlight the economy’s ability to flexibly mitigate disaster losses to a certain extent, but also reveal the vulnerability and economic disadvantage of regions that are exposed to extreme weather events. The storyline in the fourth article of my thesis investigates the interaction between global economic stress and the propagation of losses from weather extremes. I examine indirect impacts of weather extremes — tropical cyclones, heat stress, and river floods — worldwide under two different economic conditions: an unstressed economy and a globally stressed economy, as seen during the Covid-19 pandemic. I demonstrate that the adverse effects of weather extremes on global consumption are strongly amplified when the economy is under stress. Specifically, consumption losses in the USA and China double and triple, respectively, due to the global economy’s decreased capacity for disaster loss compensation. An aggravated scarcity intensifies the price response, causing consumption losses to increase. Advancing on the methods and tools used here, the final two articles in my thesis extend the agent-based model Acclimate and formalize the storyline approach. With the model extension described in the fifth article, regional consumers make rational choices on the goods bought such that their utility is maximized under a constrained budget. In an out-of-equilibrium economy, these rational consumers are shown to temporarily increase consumption of certain goods in spite of rising prices. The sixth article of my thesis proposes a formalization of the storyline framework, drawing on multiple studies including storylines presented in this thesis. The proposed guideline defines eight central elements that can be used to construct a storyline. Overall, this thesis contributes towards a better understanding of economic repercussions of weather extremes. It achieves this by providing assessments of local direct impacts, highlighting mechanisms and impacts of loss propagation, and advancing on methods and tools used. N2 - Mit dem kontinuierlichen Anstieg der globalen Mitteltemperatur aufgrund anthropogener Treibhausgasemissionen kann die Intensität und Häufigkeit vieler Wetterextreme zunehmen. Diese haben das Potential sowohl natürliche als auch menschliche Systeme stark zu beeinträchtigen. So hat die Zerstörung von Vermögenswerten und Infrastruktur sowie die Unterbrechung gesellschaftlicher und ökonomischer Abläufe oft negative wirtschaftliche Konsequenzen für direkt betroffene Regionen. Die Auswirkungen sind jedoch nicht lokal begrenzt, sondern können sich entlang von Lieferketten ausbreiten und somit auch indirekte Folgen in anderen Regionen haben – bis hin zu einer potenziell globalen wirtschaftlichen Reaktion. Daher sind Strategien zur Anpassung an veränderliche Klimabedingungen notwendig, um die Resilienz globaler Handelsketten zu stärken und dadurch negative sozioökonomische Folgen abzumildern. Hierfür ist ein besseres Verständnis lokaler Auswirkungen sowie ökonomischer Mechanismen zur Schadensausbreitung und deren Folgen erforderlich. Die vorliegende Dissertation umfasst insgesamt sechs Artikel, die zu diesem Verständnis beitragen. In diesen Studien werden zunächst lokale Auswirkungen von Wetterextremen unter gegenwärtigen und zukünftigen klimatischen Bedingungen untersucht. Weiterhin werden die globalen wirtschaftlichen Auswirkungen lokaler Wetterextreme sowie die darunterliegenden ökonomischen Effekte analysiert. In diesem Zusammenhang trägt diese Arbeit ferner zu der Weiterentwicklung der verwendeten Methoden und Ansätze bei. Der erste Artikel widmet sich zunächst der Betrachtung von extremem Schneefall in der nördlichen Hemisphäre unter dem Einfluss des Klimawandels. Zu diesem Zweck wird ein Ensemble von Projektionen globaler Klimamodelle bis zum Ende des Jahrhunderts analysiert. Die Projektionen zeigen dabei eine Verstärkung von extremen Schneefallereignissen, während die mittlere Schneefallintensität abnimmt. Um indirekte Auswirkungen von Wetterextremen zu erforschen, wird weiterhin das numerische agentenbasierte Modell Acclimate verwendet, welches die Ausbreitung ökonomischer Verlustkaskaden im globalen Versorgungsnetzwerk simuliert. In mehreren sogenannten Storylines werden die Auswirkungen eines historischen Referenzereignisses analysiert und mit den potentiellen Auswirkungen dieses Ereignisses unter plausiblen alternativen klimatischen oder sozioökonomischen Bedingungen verglichen. In dieser Dissertation werden drei zentrale Storylines vorgestellt, die jeweils unterschiedliche Aspekte der Schadensausbreitung von Wetterextremen untersuchen. Im zweiten und dritten Artikel dieser Arbeit werden dazu Storylines für die historischen Hurrikane Sandy (2012) und Harvey (2017) in den USA untersucht. Hierfür werden zunächst die lokalen ökonomischen Verluste durch diese Hurrikane ermittelt, welche als direkte wirtschaftliche Schockereignisse in Acclimate zur Berechnung der globalen Reaktion verwendet werden. Hierbei untersucht die Studie zu Hurricane Sandy globale Konsumpreisanomalien und damit einhergehende Auswirkungen auf das Konsumverhalten. Der direkte Schock löst hier eine wellenartige Veränderung globaler Konsumpreise mit drei Phasen aus, welche aus gegenläufigen Effekte aufwärts und abwärts der Lieferketten resultiert – sogenannten Upstream- und Downstream-Effekten. Zunächst steigt der Konsum aufgrund sinkender Preise durch Upstream-Effekte, bevor Preise aufgrund von Güterknappheit durch Downstream-Effekte wieder ansteigen und der Konsum abfällt. In einer Normalisierungsphase klingen diese Anomalien wieder ab. Ein länger anhaltender direkter wirtschaftlicher Schock verstärkt die Downstream-Phase und führt so in vielen Regionen insgesamt zu einem Konsumverlust. Die entwickelte Methode zur Berechnung direkter Verluste wird im dritten Artikel erweitert, indem Verstärkungen unter dem Einfluss des fortschreitenden Klimawandels berücksichtigt werden. Unter Nutzung dieser verstärkten direkten Verluste wird die Veränderung globaler Produktionsanomalien in Reaktion auf Hurricane Harvey simuliert. Die Ergebnisse zeigen, dass die USA bei zunehmender Erwärmung nicht mehr in der Lage sein werden, direkte Produktionsverluste auf nationaler Ebene auszugleichen. Stattdessen muss ein zunehmender Anteil dieser Verluste durch andere, insbesondere exportstarke Länder ausgeglichen werden. Der Anteil kleinerer Regionen an dieser ausgleichenden Produktion nimmt jedoch mit zunehmenden direkten Verlusten zu. Diese Produktionsverschiebungen verdeutlichen die Möglichkeit der globalen Wirtschaft, lokale Katastrophenverluste weitgehend flexibel abzumildern. Gleichzeitig veranschaulichen sie den Wettbewerbsnachteil direkt betroffener Wirtschaftsregionen. Die Storyline im vierten Artikel befasst sich mit dem Einfluss einer globalen wirtschaftlichen Krise auf die Schadensausbreitung von tropischen Wirbelstürmen, Hitzestress und Flussüberschwemmungen weltweit. Hierfür werden die indirekten Auswirkungen dieser Extreme unter dem Einfluss der global reduzierten wirtschaftlichen Aktivität während der Covid-19-Pandemie, sowie bei „normaler“ globaler Wirtschaftsleistung simuliert. Der Vergleich beider Szenarien zeigt bei global gestörter Wirtschaft eine deutliche Verstärkung negativer Konsumauswirkungen durch die simulierten Extreme. Konsumverluste steigen besonders stark in den USA und China an, wo sie sich verdoppeln bzw. verdreifachen. Diese Veränderungen resultieren aus der global verminderten wirtschaftlichen Kapazität, die für den Ausgleich der Produktionsverluste von Wetterextremen zur Verfügung steht. Dies verstärkt die Extremewetter-bedingte Güterknappheit, was zu Preisanstiegen und erhöhten Konsumverlusten führt. Abschließend werden in den letzten beiden Artikeln die in der Arbeit verwendeten Methoden und Ansätze erweitert. Hierfür wird das Modell Acclimate im fünften Artikel weiterentwickelt, indem Konsumenten als rational agierende Agenten modelliert werden. Mit dieser Erweiterung treffen lokale Verbraucher Entscheidungen über die konsumierten Güter so, dass diese den Nutzen eines begrenzten Budgets maximieren. Die entstehende Dynamik kann außerhalb eines wirtschaftlichen Gleichgewichts dazu führen, dass bestimmte Güter temporär trotz erhöhter Preise stärker nachgefragt werden. Der sechste Artikel formalisiert den Storyline-Ansatz und präsentiert einen Leitfaden für die Erstellung von Storylines. Dieser basiert auf den Ergebnissen mehrerer Studien, die diesen Ansatz verfolgen; einschließlich Storylines aus der vorliegenden Arbeit. Es werden insgesamt acht Elemente definiert, anhand derer eine Storyline-Studie erstellt werden kann. Insgesamt trägt diese Arbeit zu einem umfassenderen Verständnis der ökonomischen Auswirkungen von Wetterextremen bei. Hierfür werden lokale Auswirkungen von Extremen unter gegenwärtigen und zukünftigen klimatischen Bedingungen untersucht, sowie wichtige ökonomische Mechanismen und Auswirkungen der resultierenden Schadensausbreitung aufgedeckt. Neben diesen Erkenntnissen werden überdies Weiterentwicklungen der Methoden und Ansätze präsentiert, die weiterführende Analysen ermöglichen. KW - Klimawandel KW - Wetterextreme KW - indirekte ökonomische Effekte KW - makroökonomische Modellierung KW - climate change KW - weather extremes KW - indirect economic impacts KW - macro-economic modelling Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:kobv:517-opus4-611127 ER - TY - JOUR A1 - Schell, Timon A1 - Schwill, Andreas T1 - „Es ist kompliziert, alles inklusive Privatleben unter einen Hut zu bekommen“ BT - Eine Studie zu Nutzen und Schaden von Arbeitsverhältnissen für das Informatikstudium JF - Hochschuldidaktik Informatik HDI 2021 (Commentarii informaticae didacticae) N2 - Eine übliche Erzählung verknüpft lange Studienzeiten und hohe Abbrecherquoten im Informatikstudium zum einen mit der sehr gut bezahlten Nebentätigkeit von Studierenden in der Informatikbranche, die deutlich studienzeitverlängernd sei; zum anderen werde wegen des hohen Bedarfs an Informatikern ein formeller Studienabschluss von den Studierenden häufig als entbehrlich betrachtet und eine Karriere in der Informatikbranche ohne abgeschlossenes Studium begonnen. In dieser Studie, durchgeführt an der Universität Potsdam, untersuchen wir, wie viele Informatikstudierende neben dem Studium innerhalb und außerhalb der Informatikbranche arbeiten, welche Erwartungen sie neben der Bezahlung damit verbinden und wie sich die Tätigkeit auf ihr Studium und ihre spätere berufliche Perspektive auswirkt. Aus aktuellem Anlass interessieren uns auch die Auswirkungen der Covid-19-Pandemie auf die Arbeitstätigkeiten der Informatikstudierenden. KW - Informatikstudium KW - Studienabbrecher KW - Studentenjobs KW - Studiendauer Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:kobv:517-opus4-613882 SN - 978-3-86956-548-4 SN - 1868-0844 SN - 2191-1940 IS - 13 SP - 53 EP - 71 PB - Universitätsverlag Potsdam CY - Potsdam ER -