TY - GEN A1 - Mühlbauer, Felix A1 - Schröder, Lukas A1 - Skoncej, Patryk A1 - Schölzel, Mario T1 - Handling manufacturing and aging faults with software-based techniques in tiny embedded systems T2 - 18th IEEE Latin American Test Symposium (LATS 2017) N2 - Non-volatile memory area occupies a large portion of the area of a chip in an embedded system. Such memories are prone to manufacturing faults, retention faults, and aging faults. The paper presents a single software based technique that allows for handling all of these fault types in tiny embedded systems without the need for hardware support. This is beneficial for low-cost embedded systems with simple memory architectures. A software infrastructure and a flow are presented that demonstrate how the presented technique is used in general for fault handling right after manufacturing and in-the-field. Moreover, a full implementation is presented for a MSP430 microcontroller, along with a discussion of the performance, overhead, and reliability impacts. Y1 - 2027 SN - 978-1-5386-0415-1 U6 - https://doi.org/10.1109/LATW.2017.7906756 PB - IEEE CY - New York ER - TY - GEN A1 - Mühlbauer, Felix A1 - Schröder, Lukas A1 - Schölzel, Mario T1 - On hardware-based fault-handling in dynamically scheduled processors T2 - 20th International Symposium on Design and Diagnostics of Electronic Circuits & Systems (DDECS) 2017 N2 - This paper describes architectural extensions for a dynamically scheduled processor, so that it can be used in three different operation modes, ranging from high-performance, to high-reliability. With minor hardware-extensions of the control path, the resources of the superscalar data-path can be used either for high-performance execution, fail-safe-operation, or fault-tolerant-operation. This makes the processor-architecture a very good candidate for applications with dynamically changing reliability requirements, e.g. for automotive applications. The paper reports the hardware-overhead for the extensions, and investigates the performance penalties introduced by the fail-safe and fault-tolerant mode. Furthermore, a comprehensive fault simulation was carried out in order to investigate the fault-coverage of the proposed approach. Y1 - 2017 SN - 978-1-5386-0472-4 U6 - https://doi.org/10.1109/DDECS.2017.7934572 SN - 2334-3133 SN - 2473-2117 SP - 201 EP - 206 PB - IEEE CY - New York ER - TY - GEN A1 - Fabian, Benjamin A1 - Baumann, Annika A1 - Ehlert, Mathias A1 - Ververis, Vasilis A1 - Ermakova, Tatiana T1 - CORIA - Analyzing internet connectivity risks using network graphs T2 - 2017 IEEE International Conference on Communications (ICC) N2 - The Internet can be considered as the most important infrastructure for modern society and businesses. A loss of Internet connectivity has strong negative financial impacts for businesses and economies. Therefore, assessing Internet connectivity, in particular beyond their own premises and area of direct control, is of growing importance in the face of potential failures, accidents, and malicious attacks. This paper presents CORIA, a software framework for an easy analysis of connectivity risks based on large network graphs. It provides researchers, risk analysts, network managers and security consultants with a tool to assess an organization's connectivity and paths options through the Internet backbone, including a user-friendly and insightful visual representation of results. CORIA is flexibly extensible in terms of novel data sets, graph metrics, and risk scores that enable further use cases. The performance of CORIA is evaluated by several experiments on the Internet graph and further randomly generated networks. KW - risk analysis KW - connectivity KW - graph analysis KW - complex networks KW - Internet Y1 - 2017 SN - 978-1-4673-8999-0 SN - 978-1-4673-9000-2 U6 - https://doi.org/10.1109/ICC.2017.7996828 SN - 1550-3607 PB - IEEE CY - Piscataway ER - TY - JOUR A1 - Gebser, Martin A1 - Maratea, Marco A1 - Ricca, Francesco T1 - The sixth answer set programming competition JF - Journal of artificial intelligence research : JAIR N2 - Answer Set Programming (ASP) is a well-known paradigm of declarative programming with roots in logic programming and non-monotonic reasoning. Similar to other closely related problemsolving technologies, such as SAT/SMT, QBF, Planning and Scheduling, advancements in ASP solving are assessed in competition events. In this paper, we report about the design and results of the Sixth ASP Competition, which was jointly organized by the University of Calabria (Italy), Aalto University (Finland), and the University of Genoa (Italy), in affiliation with the 13th International Conference on Logic Programming and Non-Monotonic Reasoning. This edition maintained some of the design decisions introduced in 2014, e.g., the conception of sub-tracks, the scoring scheme,and the adherence to a fixed modeling language in order to push the adoption of the ASP-Core-2 standard. On the other hand, it featured also some novelties, like a benchmark selection stage classifying instances according to their empirical hardness, and a “Marathon” track where the topperforming systems are given more time for solving hard benchmarks. Y1 - 2017 U6 - https://doi.org/10.1613/jair.5373 SN - 1076-9757 SN - 1943-5037 VL - 60 SP - 41 EP - 95 PB - AI Access Found. CY - Marina del Rey ER - TY - GEN A1 - Saint-Dizier, Patrick A1 - Stede, Manfred T1 - Foundations of the language of argumentation T2 - Argument & computation Y1 - 2017 U6 - https://doi.org/10.3233/AAC-170018 SN - 1946-2166 SN - 1946-2174 VL - 8 IS - 2 Special issue SP - 91 EP - 93 PB - IOS Press CY - Amsterdam ER - TY - GEN A1 - Giese, Holger ED - Kouchnarenko, Olga ED - Khosravi, Ramtin T1 - Formal models and analysis for self-adaptive cyber-physical systems BT - (extended abstract) T2 - Lecture notes in computer science N2 - In this extended abstract, we will analyze the current challenges for the envisioned Self-Adaptive CPS. In addition, we will outline our results to approach these challenges with SMARTSOS [10] a generic approach based on extensions of graph transformation systems employing open and adaptive collaborations and models at runtime for trustworthy self-adaptation, self-organization, and evolution of the individual systems and the system-of-systems level taking the independent development, operation, management, and evolution of these systems into account. Y1 - 2017 SN - 978-3-319-57666-4 SN - 978-3-319-57665-7 U6 - https://doi.org/10.1007/978-3-319-57666-4_1 SN - 0302-9743 SN - 1611-3349 VL - 10231 SP - 3 EP - 9 PB - Springer CY - Cham ER - TY - JOUR A1 - Lindauer, Marius A1 - Hoos, Holger A1 - Leyton-Brown, Kevin A1 - Schaub, Torsten T1 - Automatic construction of parallel portfolios via algorithm configuration JF - Artificial intelligence N2 - Since 2004, increases in computational power described by Moore's law have substantially been realized in the form of additional cores rather than through faster clock speeds. To make effective use of modern hardware when solving hard computational problems, it is therefore necessary to employ parallel solution strategies. In this work, we demonstrate how effective parallel solvers for propositional satisfiability (SAT), one of the most widely studied NP-complete problems, can be produced automatically from any existing sequential, highly parametric SAT solver. Our Automatic Construction of Parallel Portfolios (ACPP) approach uses an automatic algorithm configuration procedure to identify a set of configurations that perform well when executed in parallel. Applied to two prominent SAT solvers, Lingeling and clasp, our ACPP procedure identified 8-core solvers that significantly outperformed their sequential counterparts on a diverse set of instances from the application and hard combinatorial category of the 2012 SAT Challenge. We further extended our ACPP approach to produce parallel portfolio solvers consisting of several different solvers by combining their configuration spaces. Applied to the component solvers of the 2012 SAT Challenge gold medal winning SAT Solver pfolioUZK, our ACPP procedures produced a significantly better-performing parallel SAT solver. KW - Algorithm configuration KW - Parallel SAT solving KW - Algorithm portfolios KW - Programming by optimization KW - Automated parallelization Y1 - 2016 U6 - https://doi.org/10.1016/j.artint.2016.05.004 SN - 0004-3702 SN - 1872-7921 VL - 244 SP - 272 EP - 290 PB - Elsevier CY - Amsterdam ER - TY - JOUR A1 - Baedke, Jan A1 - Schöttler, Tobias T1 - Visual Metaphors in the Sciences BT - the Case of Epigenetic Landscape Images JF - Journal for General Philosophy of Science N2 - Recent philosophical analyses of the epistemic dimension of images in the sciences show a certain trend in acknowledging potential roles of these images beyond their merely decorative or pedagogical functions. We argue, however, that this new debate has yet paid little attention to a special type of pictures, we call ‘visual metaphor’, and its versatile heuristic potential in organizing data, supporting communication, and guiding research, modeling, and theory formation. Based on a case study of Conrad Hal Waddington’s epigenetic landscape images in biology, we develop a descriptive framework applicable to heuristic roles of various visual metaphors in the sciences. KW - Conrad Hal Waddington KW - Epigenetic landscape KW - Modelling KW - Scientific images KW - Theory formation KW - Visual metaphor Y1 - 2017 U6 - https://doi.org/10.1007/s10838-016-9353-9 SN - 0925-4560 SN - 1572-8587 VL - 48 SP - 173 EP - 194 PB - Springer CY - Dordrecht ER - TY - JOUR A1 - Bailis, Peter A1 - Dillahunt, Tawanna A1 - Müller, Stefanie A1 - Baudisch, Patrick T1 - Research for Practice: Technology for Underserved Communities; Personal Fabrication JF - Communications of the ACM / Association for Computing Machinery N2 - THIS INSTALLMENT OF Research for Practice provides curated reading guides to technology for underserved communities and to new developments in personal fabrication. First, Tawanna Dillahunt describes design considerations and technology for underserved and impoverished communities. Designing for the more than 1.6 billion impoverished individuals worldwide requires special consideration of community needs, constraints, and context. Her selections span protocols for poor-quality communication networks, community-driven content generation, and resource and public service discovery. Second, Stefanie Mueller and Patrick Baudisch provide an overview of recent advances in personal fabrication (for example, 3D printers). Y1 - 2017 U6 - https://doi.org/10.1145/3080188 SN - 0001-0782 SN - 1557-7317 VL - 60 SP - 46 EP - 49 PB - Association for Computing Machinery CY - New York ER - TY - CHAP A1 - Kiy, Alexander A1 - List, Christoph A1 - Lucke, Ulrike T1 - A virtual environment and infrastructure to ensure future readiness of Computer Centers T2 - European Journal of Higher Education IT N2 - The ongoing digitalization leads to a need of continuous change of ICT (Information and Communi-cation Technology) in all university domains and therefore affects all stakeholders in this arena. More and more ICT components, systems and tools occur and have to be integrated into the existing processes and infrastructure of the institutions. These tasks include the transfer of resources and information across multiple ICT systems. By using so-called virtual environments for domains of re-search, education, learning and work, the performance of daily tasks can be aided. Based on a user requirement analysis different short- and long-term objectives were identified and are tackled now in the context of a federal research project. In order to be prepared for the ongoing digitalization, new systems have to be provided. Both, a service-oriented infrastructure and a related web-based virtual learning environment constitute the platform Campus.UP and creates the necessary basis to be ready for future challenges. The current focus lies on e-portfolio work, hence we will present a related focus group evaluation. The results indicate a tremendous need to extend the possibilities of sharing resources across system boundaries, in order to enable a comfortable participation of exter-nal cooperating parties and to clarify the focus of each connected system. The introduction of such an infrastructure implies far-reaching changes for traditional data centers. Therefore, the challenges and risks of faculty conducting innovation projects for the ICT organization are taken as a starting point to stimulate a discussion, how data centers can utilize projects to be ready for the future needs. We are confident that Campus.UP will provide the basis for ensuring the persistent transfer of innovation to the ICT organization and thus will contribute to tackle the future challenges of digitalization. Y1 - 2017 UR - http://www.eunis.org/erai/2017-1/ SN - 2519-1764 VL - 2017 IS - 1 ER - TY - THES A1 - Christgau, Steffen T1 - One-sided communication on a non-cache-coherent many-core architecture T1 - Einseitige Kommunikation auf einer nicht-cache-kohärenten Vielkern-Prozessorarchitektur N2 - Aktuelle Mehrkernprozessoren stellen parallele Systeme dar, die den darauf ausgeführten Programmen gemeinsamen Speicher zur Verfügung stellen. Sowohl die ansteigende Kernanzahlen in sogenannten Vielkernprozessoren (many-core processors) als auch die weiterhin steigende Leistungsfähigkeit der einzelnen Kerne erfordert hohe Bandbreiten, die das Speichersystem des Prozessors liefern muss. Hardware-basierte Cache-Kohärenz stößt in aktuellen Vielkernprozessoren an Grenzen des praktisch Machbaren. Dementsprechend müssen alternative Architekturen und entsprechend geeignete Programmiermodelle untersucht werden. In dieser Arbeit wird der Single-Chip Cloud Computer (SCC), ein nicht-cachekohärenter Vielkernprozessor betrachtet, der aus 48, über ein Gitternetzwerk verbundenen Kernen besteht. Obwohl der Prozessor für nachrichten-basierte Kommunikation entwickelt worden ist, zeigen die Ergebnisse dieser Arbeit, dass einseitige Kommunikation auf Basis gemeinsamen Speichers effizient auf diesem Architekturtyp realisiert werden kann. Einseitige Kommunikation ermöglicht Datenaustausch zwischen Prozessen, bei der der Empfänger keine Details über die stattfindende Kommunikation besitzen muss. Im Sinne des MPI-Standards ist so ein Zugriff auf Speicher entfernter Prozesse möglich. Zur Umsetzung dieses Konzepts auf nicht-kohärenten Architekturen werden in dieser Arbeit sowohl eine effiziente Prozesssynchronisation als auch ein Kommunikationsschema auf Basis von software-basierter Cache-Kohärenz erarbeitet und untersucht. Die Prozesssynchronisation setzt das Konzept der general active target synchronization aus dem MPI-Standard um. Ein existierendes Klassifikationsschema für dessen Implementierungen wird erweitert und zur Identifikation einer geeigneten Klasse für die nicht-kohärente Plattform des SCC verwendet. Auf Grundlage der Klassifikation werden existierende Implementierungen analysiert, daraus geeignete Konzepte extrahiert und ein leichtgewichtiges Synchronisationsprotokoll für den SCC entwickelt, das sowohl gemeinsamen Speicher als auch ungecachete Speicherzugriffe verwendet. Das vorgestellte Schema ist nicht anfällig für Verzögerungen zwischen Prozessen und erlaubt direkte Kommunikation sobald beide Kommunikationspartner dafür bereit sind. Die experimentellen Ergebnisse zeigen ein sehr gutes Skaliserungsverhalten und eine fünffach geringere Latenz für die Prozesssynchronisation im Vergleich zu einer auf Nachrichten basierenden MPI-Implementierung des SCC. Für die Kommunikation wird mit SCOSCo ein auf gemeinsamen Speicher und software-basierter Cache-Kohärenz basierenden Konzept vorgestellt. Entsprechende Anforderungen an die Kohärenz, die dem MPI-Standard entsprechen, werden aufgestellt und eine schlanke Implementierung auf Basis der Hard- und Software-Funktionalitäten des SCCs entwickelt. Trotz einer aufgedecktem Fehlfunktion im Speichersubsystem des SCC kann in den experimentellen Auswertungen von Mikrobenchmarks eine fünffach verbesserte Bandbreite und eine nahezu vierfach verringerte Latenz beobachtet werden. In Anwendungsexperimenten, wie einer dreidimensionalen schnellen Fourier-Transformation, kann der Anteil der Kommunikation an der Laufzeit um den Faktor fünf reduziert werden. In Ergänzung dazu werden in dieser Arbeit Konzepte aufgestellt, die in zukünftigen Architekturen, die Cache-Kohärenz nicht auf einer globalen Ebene des Prozessors liefern können, für die Umsetzung von Software-basierter Kohärenz für einseitige Kommunikation hilfreich sind. N2 - Contemporary multi-core processors are parallel systems that also provide shared memory for programs running on them. Both the increasing number of cores in so-called many-core systems and the still growing computational power of the cores demand for memory systems that are able to deliver high bandwidths. Caches are essential components to satisfy this requirement. Nevertheless, hardware-based cache coherence in many-core chips faces practical limits to provide both coherence and high memory bandwidths. In addition, a shift away from global coherence can be observed. As a result, alternative architectures and suitable programming models need to be investigated. This thesis focuses on fast communication for non-cache-coherent many-core architectures. Experiments are conducted on the Single-Chip Cloud Computer (SCC), a non-cache-coherent many-core processor with 48 mesh-connected cores. Although originally designed for message passing, the results of this thesis show that shared memory can be efficiently used for one-sided communication on this kind of architecture. One-sided communication enables data exchanges between processes where the receiver is not required to know the details of the performed communication. In the notion of the Message Passing Interface (MPI) standard, this type of communication allows to access memory of remote processes. In order to support this communication scheme on non-cache-coherent architectures, both an efficient process synchronization and a communication scheme with software-managed cache coherence are designed and investigated. The process synchronization realizes the concept of the general active target synchronization scheme from the MPI standard. An existing classification of implementation approaches is extended and used to identify an appropriate class for the non-cache-coherent shared memory platform. Based on this classification, existing implementations are surveyed in order to find beneficial concepts, which are then used to design a lightweight synchronization protocol for the SCC that uses shared memory and uncached memory accesses. The proposed scheme is not prone to process skew and also enables direct communication as soon as both communication partners are ready. Experimental results show very good scaling properties and up to five times lower synchronization latency compared to a tuned message-based MPI implementation for the SCC. For the communication, SCOSCo, a shared memory approach with software-managed cache coherence, is presented. According requirements for the coherence that fulfill MPI's separate memory model are formulated, and a lightweight implementation exploiting SCC hard- and software features is developed. Despite a discovered malfunction in the SCC's memory subsystem, the experimental evaluation of the design reveals up to five times better bandwidths and nearly four times lower latencies in micro-benchmarks compared to the SCC-tuned but message-based MPI library. For application benchmarks, like a parallel 3D fast Fourier transform, the runtime share of communication can be reduced by a factor of up to five. In addition, this thesis postulates beneficial hardware concepts that would support software-managed coherence for one-sided communication on future non-cache-coherent architectures where coherence might be only available in local subdomains but not on a global processor level. KW - parallel programming KW - middleware KW - Message Passing Interface KW - one-sided communication KW - software-based cache coherence KW - process synchronization KW - parallele Programmierung KW - einseitige Kommunikation KW - Software-basierte Cache-Kohärenz KW - Prozesssynchronisierung KW - Message Passing Interface KW - Middleware Y1 - 2017 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:kobv:517-opus4-403100 ER - TY - THES A1 - Al-Areqi, Samih Taha Mohammed T1 - Semantics-based automatic geospatial service composition T1 - Semantikbasierte automatische Komposition von GIS-Diensten N2 - Although it has become common practice to build applications based on the reuse of existing components or services, technical complexity and semantic challenges constitute barriers to ensuring a successful and wide reuse of components and services. In the geospatial application domain, the barriers are self-evident due to heterogeneous geographic data, a lack of interoperability and complex analysis processes. Constructing workflows manually and discovering proper services and data that match user intents and preferences is difficult and time-consuming especially for users who are not trained in software development. Furthermore, considering the multi-objective nature of environmental modeling for the assessment of climate change impacts and the various types of geospatial data (e.g., formats, scales, and georeferencing systems) increases the complexity challenges. Automatic service composition approaches that provide semantics-based assistance in the process of workflow design have proven to be a solution to overcome these challenges and have become a frequent demand especially by end users who are not IT experts. In this light, the major contributions of this thesis are: (i) Simplification of service reuse and workflow design of applications for climate impact analysis by following the eXtreme Model-Driven Development (XMDD) paradigm. (ii) Design of a semantic domain model for climate impact analysis applications that comprises specifically designed services, ontologies that provide domain-specific vocabulary for referring to types and services, and the input/output annotation of the services using the terms defined in the ontologies. (iii) Application of a constraint-driven method for the automatic composition of workflows for analyzing the impacts of sea-level rise. The application scenario demonstrates the impact of domain modeling decisions on the results and the performance of the synthesis algorithm. N2 - Obwohl es gängige Praxis geworden ist, Anwendungen basierend auf der Wiederverwendung von existierenden Komponenten oder Diensten zu bauen, stellen technische Komplexität und semantische Herausforderungen Hindernisse beim Sicherstellen einer erfolgreichen und breiten Wiederverwendungen von Komponenten und Diensten. In der geowissenschaftlichen Anwendungsdomäne sind die Hindernisse durch heterogene geografische Daten, fehlende Interoperabilität und komplexe Analyseprozessen besonders offensichtlich. Workflows manuell zu konstruieren und passende Dienste und Daten zu finden, welche die Nutzerabsichten und -präferenzen abdecken, ist schwierig und zeitaufwändig besonders für Nutzer, die nicht in der Softwareentwicklung ausgebildet sind. Zudem erhöhen die verschiedenen Zielrichtungen der Umweltmodellierung für die Bewertung der Auswirkungen von Klimaänderungen und die unterschiedlichen Typen geografischer Daten (z.B. Formate, Skalierungen, und Georeferenzsysteme) die Komplexität. Automatische Dienstkompositionsansätze, die Semantik-basierte Unterstützung im Prozess des Workflowdesigns zur Verfügung stellen, haben bewiesen eine Lösung zur Bewältigung dieser Herausforderungen zu sein und sind besonders von Endnutzern, die keine IT-Experten sind, eine häufige Forderung geworden. Unter diesem Gesichtspunkt sind die Hauptbeiträge dieser Doktorarbeit: I. Vereinfachung der Wiederverwendung von Diensten und des Workflowdesigns von Klimafolgenanalysen durch Anwendung des Paradigma des eXtreme Model-Driven Development (XMDD) II. Design eines semantischen Domänenmodells für Anwendungen der Klimafolgenanalysen, welches speziell entwickelte Dienste, Ontologien (die domänen-spezifisches Vokabular zur Verfügung stellen, um Typen und Dienste zu beschreiben), und Eingabe-/Ausgabe-Annotationen der Dienste (unter Verwendung von Begriffen, die in den Ontologien definiert sind) enthält. III. Anwendungen einer Constraint-getriebenen Methode für die automatische Komposition von Workflows zum Analysieren der Auswirkungen des Meeresspiegelanstiegs. Das Anwendungsszenario demonstriert die Auswirkung von Domänenmodellierungsentscheidungen auf die Ergebnisse und die Laufzeit des Synthesealgorithmus. KW - geospatial services KW - service composition KW - scientific workflows KW - semantic domain modeling KW - ontologies KW - climate impact analysis KW - GIS-Dienstkomposition KW - Wissenschaftlichesworkflows KW - semantische Domänenmodellierung KW - Ontologien KW - Klimafolgenanalyse Y1 - 2017 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:kobv:517-opus4-402616 ER - TY - THES A1 - Felgentreff, Tim T1 - The Design and Implementation of Object-Constraint Programming Y1 - 2017 ER -