004 Datenverarbeitung; Informatik
Refine
Year of publication
- 2014 (49) (remove)
Document Type
- Article (19)
- Doctoral Thesis (15)
- Monograph/Edited Volume (8)
- Postprint (5)
- Master's Thesis (1)
- Preprint (1)
Keywords
- cloud computing (4)
- Cloud Computing (3)
- Forschungsprojekte (2)
- Future SOC Lab (2)
- Geschäftsprozessmanagement (2)
- In-Memory Technologie (2)
- Multicore Architekturen (2)
- RDF (2)
- business process management (2)
- multicore architectures (2)
- research projects (2)
- 47A52 (1)
- 65R20 (1)
- 65R32 (1)
- 78A46 (1)
- Algorithmenablaufplanung (1)
- Algorithmenkonfiguration (1)
- Algorithmenselektion (1)
- Anfragepaare (1)
- Antwortmengenprogrammierung (1)
- Assoziationsregeln (1)
- Basic Storage Anbieter (1)
- Bayesian networks (1)
- Bayessche Netze (1)
- Betriebssysteme (1)
- Bitcoin (1)
- Boolean constraint solver (1)
- Carrera Digital D132 (1)
- Cloud (1)
- Cloud Datenzentren (1)
- Conformance Überprüfung (1)
- Datenflusskorrektheit (1)
- Datenintegration (1)
- Datenreinigung (1)
- Datenvertraulichkeit (1)
- Differential Privacy (1)
- E-Learning (1)
- Echtzeit (1)
- Energieeffizienz (1)
- Entwicklungswerkzeuge (1)
- Erfahrungsbericht (1)
- Fehlende Daten (1)
- Forschungskolleg (1)
- Fredholm complexes (1)
- HCI (1)
- Hasso Plattner Institute (1)
- Hasso-Plattner-Institut (1)
- Hauptspeicher Technologie (1)
- Homomorphe Verschlüsselung (1)
- In-Memory technology (1)
- Informationsvorhaltung (1)
- Klausurtagung (1)
- LEGO Mindstorms EV3 (1)
- LOD (1)
- Laufzeitverhalten (1)
- Lefschetz number (1)
- MOOCs (1)
- Mensch-Computer-Interaktion (1)
- Mustererkennung (1)
- Objektlebenszyklus-Synchronisation (1)
- Online-Lernen (1)
- Onlinekurs (1)
- Petri net Mapping (1)
- Petri net mapping (1)
- Ph.D. Retreat (1)
- Privacy (1)
- Probabilistische Modelle (1)
- Process Mining (1)
- RT_PREEMT patch (1)
- RT_PREEMT-Patch (1)
- Research School (1)
- Ressourcenmanagement (1)
- SPARQL (1)
- Service-oriented Systems Engineering (1)
- Sicherheit (1)
- Softwareanalyse (1)
- Softwareentwicklung (1)
- Softwareentwicklungsprozesse (1)
- Softwaretechnik (1)
- Softwarevisualisierung (1)
- Softwarewartung (1)
- Synonyme (1)
- Systembiologie (1)
- Tele-Lab (1)
- Tele-Teaching (1)
- Testen (1)
- Threshold Cryptography (1)
- Vernetzte Daten (1)
- Versionierung (1)
- Visualisierung (1)
- Vorhersage (1)
- aerosol size distribution (1)
- algorithm configuration (1)
- algorithm scheduling (1)
- algorithm selection (1)
- answer set programming (1)
- assistive Technologien (1)
- assistive technologies (1)
- association rule mining (1)
- automotive electronics (1)
- basic cloud storage services (1)
- bitcoin (1)
- categories (1)
- changeability (1)
- cleansing (1)
- cloud (1)
- cloud datacenter (1)
- computer science (1)
- confidentiality (1)
- confluence (1)
- conformance checking (1)
- data (1)
- data flow correctness (1)
- database technology (1)
- depressive symptoms (1)
- development tools (1)
- differential privacy (1)
- dimensional (1)
- distributed systems (1)
- dynamic consolidation (1)
- dynamische Umsortierung (1)
- e-learning (1)
- eingebettete Systeme (1)
- elliptic complexes (1)
- embedded systems (1)
- empirical studies (1)
- empirische Studien (1)
- energy efficiency (1)
- experience report (1)
- fMRI (1)
- future SOC lab (1)
- ganzheitlich (1)
- geovisualization (1)
- grammars (1)
- graph-transformations (1)
- holistic (1)
- homomorphic encryption (1)
- in-memory technology (1)
- integral equation (1)
- inverse ill-posed problem (1)
- inverse scattering (1)
- iterative regularization (1)
- knowledge discovery (1)
- laser remote sensing (1)
- level-replacement systems (1)
- linked data (1)
- logical signaling networks (1)
- logische Signalnetzwerke (1)
- main memory computing (1)
- map reduce (1)
- missing data (1)
- model (1)
- model-driven engineering (1)
- modelgetriebene Entwicklung (1)
- monetary incentive delay task (1)
- object life cycle synchronization (1)
- online assistance (1)
- online course (1)
- online-learning (1)
- open source software (1)
- openHPI (1)
- operating systems (1)
- parallel (1)
- parallel solving (1)
- paralleles Lösen (1)
- prediction (1)
- prefetching (1)
- privacy (1)
- probabilistic models (1)
- process mining (1)
- programs (1)
- public cloud storage services (1)
- query matching (1)
- real-time (1)
- resource management (1)
- reward system (1)
- runtime behavior (1)
- sat (1)
- security (1)
- sign language (1)
- software analysis (1)
- software development (1)
- software development processes (1)
- software engineering (1)
- software maintenance (1)
- software visualization (1)
- solver (1)
- spatio-temporal sensor data (1)
- stochastic Petri nets (1)
- stochastische Petri Netze (1)
- synchronization (1)
- synonym discovery (1)
- systems biology (1)
- tele-TASK (1)
- tele-lab (1)
- tele-teaching (1)
- testing (1)
- threshold cryptography (1)
- user interfaces (1)
- ventral striatum (1)
- versioning (1)
- verteilte Datenbanken (1)
- virtualisierte IT-Infrastruktur (1)
- visualization (1)
- Änderbarkeit (1)
- öffentliche Cloud Speicherdienste (1)
Institute
- Institut für Informatik und Computational Science (25)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (17)
- Mathematisch-Naturwissenschaftliche Fakultät (4)
- Department Sport- und Gesundheitswissenschaften (1)
- Extern (1)
- Hasso-Plattner-Institut für Digital Engineering GmbH (1)
- Institut für Mathematik (1)
- Institut für Umweltwissenschaften und Geographie (1)
This book presents an agile and model-driven approach to manage scientific workflows. The approach is based on the Extreme Model Driven Design (XMDD) paradigm and aims at simplifying and automating the complex data analysis processes carried out by scientists in their day-to-day work. Besides documenting the impact the workflow modeling might have on the work of natural scientists, this book serves three major purposes: 1. It acts as a primer for practitioners who are interested to learn how to think in terms of services and workflows when facing domain-specific scientific processes. 2. It provides interesting material for readers already familiar with this kind of tools, because it introduces systematically both the technologies used in each case study and the basic concepts behind them. 3. As the addressed thematic field becomes increasingly relevant for lectures in both computer science and experimental sciences, it also provides helpful material for teachers that plan similar courses.
Geometric generalization is a fundamental concept in the digital mapping process. An increasing amount of spatial data is provided on the web as well as a range of tools to process it. This jABC workflow is used for the automatic testing of web-based generalization services like mapshaper.org by executing its functionality, overlaying both datasets before and after the transformation and displaying them visually in a .tif file. Mostly Web Services and command line tools are used to build an environment where ESRI shapefiles can be uploaded, processed through a chosen generalization service and finally visualized in Irfanview.
In the geoinformatics field, remote sensing data is often used for analyzing the characteristics of the current investigation area. This includes DEMs, which are simple raster grids containing grey scales representing the respective elevation values. The project CREADED that is presented in this paper aims at making these monochrome raster images more significant and more intuitively interpretable. For this purpose, an executable interactive model for creating a colored and relief-shaded Digital Elevation Model (DEM) has been designed using the jABC framework. The process is based on standard jABC-SIBs and SIBs that provide specific GIS functions, which are available as Web services, command line tools and scripts.
This paper describes the implementation of a workflow model for service-oriented computing of potential areas for wind turbines in jABC. By implementing a re-executable model the manual effort of a multi-criteria site analysis can be reduced. The aim is to determine the shift of typical geoprocessing tools of geographic information systems (GIS) from the desktop to the web. The analysis is based on a vector data set and mainly uses web services of the “Center for Spatial Information Science and Systems” (CSISS). This paper discusses effort, benefits and problems associated with the use of the web services.
Location analyses are among the most common tasks while working with spatial data and geographic information systems. Automating the most frequently used procedures is therefore an important aspect of improving their usability. In this context, this project aims to design and implement a workflow, providing some basic tools for a location analysis. For the implementation with jABC, the workflow was applied to the problem of finding a suitable location for placing an artificial reef. For this analysis three parameters (bathymetry, slope and grain size of the ground material) were taken into account, processed, and visualized with the The Generic Mapping Tools (GMT), which were integrated into the workflow as jETI-SIBs. The implemented workflow thereby showed that the approach to combine jABC with GMT resulted in an user-centric yet user-friendly tool with high-quality cartographic outputs.
Creation of topographic maps
(2014)
Location analyses are among the most common tasks while working with spatial data and geographic information systems. Automating the most frequently used procedures is therefore an important aspect of improving their usability. In this context, this project aims to design and implement a workflow, providing some basic tools for a location analysis. For the implementation with jABC, the workflow was applied to the problem of finding a suitable location for placing an artificial reef. For this analysis three parameters (bathymetry, slope and grain size of the ground material) were taken into account, processed, and visualized with the The Generic Mapping Tools (GMT), which were integrated into the workflow as jETI-SIBs. The implemented workflow thereby showed that the approach to combine jABC with GMT resulted in an user-centric yet user-friendly tool with high-quality cartographic outputs.
GraffDok is an application helping to maintain an overview over sprayed images somewhere in a city. At the time of writing it aims at vandalism rather than at beautiful photographic graffiti in an underpass. Looking at hundreds of tags and scribbles on monuments, house walls, etc. it would be interesting to not only record them in writing but even make them accessible electronically, including images.
GraffDok’s workflow is simple and only requires an EXIF-GPS-tagged photograph of a graffito. It automatically determines its location by using reverse geocoding with the given GPS-coordinates and the Gisgraphy WebService. While asking the user for some more meta data, GraffDok analyses the image in parallel with this and tries to detect fore- and background – before extracting the drawing lines and make them stand alone. The command line based tool ImageMagick is used here as well as for accessing EXIF data.
Any meta data is written to csv-files, which will stay easily accessible and can be integrated in TeX-files as well. The latter ones are converted to PDF at the end of the workflow, containing a table about all graffiti and a summary for each – including the generated characteristic graffiti pattern image.
The data quality of real-world datasets need to be constantly monitored and maintained to allow organizations and individuals to reliably use their data. Especially, data integration projects suffer from poor initial data quality and as a consequence consume more effort and money. Commercial products and research prototypes for data cleansing and integration help users to improve the quality of individual and combined datasets. They can be divided into either standalone systems or database management system (DBMS) extensions. On the one hand, standalone systems do not interact well with DBMS and require time-consuming data imports and exports. On the other hand, DBMS extensions are often limited by the underlying system and do not cover the full set of data cleansing and integration tasks.
We overcome both limitations by implementing a concise set of five data cleansing and integration operators on the parallel data analytics platform Stratosphere. We define the semantics of the operators, present their parallel implementation, and devise optimization techniques for individual operators and combinations thereof. Users specify declarative queries in our query language METEOR with our new operators to improve the data quality of individual datasets or integrate them to larger datasets. By integrating the data cleansing operators into the higher level language layer of Stratosphere, users can easily combine cleansing operators with operators from other domains, such as information extraction, to complex data flows. Through a generic description of the operators, the Stratosphere optimizer reorders operators even from different domains to find better query plans.
As a case study, we reimplemented a part of the large Open Government Data integration project GovWILD with our new operators and show that our queries run significantly faster than the original GovWILD queries, which rely on relational operators. Evaluation reveals that our operators exhibit good scalability on up to 100 cores, so that even larger inputs can be efficiently processed by scaling out to more machines. Finally, our scripts are considerably shorter than the original GovWILD scripts, which results in better maintainability of the scripts.
Process models specify behavioral execution constraints between activities as well as between activities and data objects. A data object is characterized by its states and state transitions represented as object life cycle. For process execution, all behavioral execution constraints must be correct. Correctness can be verified via soundness checking which currently only considers control flow information. For data correctness, conformance between a process model and its object life cycles is checked. Current approaches abstract from dependencies between multiple data objects and require fully specified process models although, in real-world process repositories, often underspecified models are found. Coping with these issues, we introduce the concept of synchronized object life cycles and we define a mapping of data constraints of a process model to Petri nets extending an existing mapping. Further, we apply the notion of weak conformance to process models to tell whether each time an activity needs to access a data object in a particular state, it is guaranteed that the data object is in or can reach the expected state. Then, we introduce an algorithm for an integrated verification of control flow correctness and weak data conformance using soundness checking.
Cloud-RAID
(2014)