Filtern
Erscheinungsjahr
Dokumenttyp
- Monographie/Sammelband (89) (entfernen)
Sprache
- Englisch (89) (entfernen)
Schlagworte
- Hasso-Plattner-Institut (7)
- Forschungskolleg (6)
- Hasso Plattner Institute (6)
- Klausurtagung (6)
- Service-oriented Systems Engineering (6)
- Forschungsprojekte (4)
- Future SOC Lab (4)
- In-Memory Technologie (4)
- Modellierung (4)
- Multicore Architekturen (4)
- Research School (4)
- cloud computing (4)
- Cloud Computing (3)
- Datenintegration (3)
- Graphtransformationen (3)
- Model Synchronisation (3)
- Model Transformation (3)
- Ph.D. Retreat (3)
- Ph.D. retreat (3)
- Tripel-Graph-Grammatik (3)
- Verifikation (3)
- Virtualisierung (3)
- graph transformation (3)
- multicore architectures (3)
- research projects (3)
- service-oriented systems engineering (3)
- virtual machines (3)
- AUTOSAR (2)
- BPMN (2)
- Cloud-Sicherheit (2)
- Cloud-Speicher (2)
- Data Integration (2)
- Graphtransformationssysteme (2)
- In-Memory technology (2)
- Model Synchronization (2)
- Model-Driven Engineering (2)
- Modeling (2)
- Privacy (2)
- Process Modeling (2)
- Prozessmodellierung (2)
- Ressourcenoptimierung (2)
- Sicherheit (2)
- SysML (2)
- Virtuelle Maschinen (2)
- data profiling (2)
- graph constraints (2)
- in-memory technology (2)
- incremental graph pattern matching (2)
- modeling (2)
- modellgetriebene Entwicklung (2)
- nested graph conditions (2)
- privacy (2)
- research school (2)
- security (2)
- verification (2)
- virtualization (2)
- virtuelle Maschinen (2)
- Abhängigkeiten (1)
- Abstraktion von Geschäftsprozessmodellen (1)
- Apriori (1)
- Architektur (1)
- Aspect-oriented Programming (1)
- Aspektorientierte Softwareentwicklung (1)
- Association Rule Mining (1)
- Assoziationsregeln (1)
- Asynchrone Schaltung (1)
- Asynchronous circuit (1)
- Attribut-Merge-Prozess (1)
- Attribute Merge Process (1)
- Ausführung von Modellen (1)
- BPM (1)
- Batchprozesse (1)
- Bayes'sche Netze (1)
- Bayesian networks (1)
- Bedingte Inklusionsabhängigkeiten (1)
- Beschränkungen und Abhängigkeiten (1)
- Betriebssysteme (1)
- Bisimulation (1)
- Bitcoin (1)
- CEP (1)
- CSC (1)
- CSCW (1)
- Change Management (1)
- Cloud (1)
- Conditional Inclusion Dependency (1)
- Conformance Überprüfung (1)
- Constraints (1)
- Context-oriented Programming (1)
- Contracts (1)
- Controller-Resynthese (1)
- Cyber-Physical Systems (1)
- Cyber-Physical-Systeme (1)
- Cyber-physical-systems (1)
- Data Dependency (1)
- Data Modeling (1)
- Data Profiling (1)
- Data Quality (1)
- Data Warehouse (1)
- Database Cost Model (1)
- Datenabhängigkeiten (1)
- Datenanalyse (1)
- Datenbank-Kostenmodell (1)
- Datenflusskorrektheit (1)
- Datenmodellierung (1)
- Datenqualität (1)
- Datenvertraulichkeit (1)
- Deadline-Verbreitung (1)
- Design Thinking (1)
- Differential Privacy (1)
- Discrimination Networks (1)
- Duplicate Detection (1)
- Duplikaterkennung (1)
- Dynamic Type System (1)
- Dynamische Typ Systeme (1)
- E-Learning (1)
- EHR (1)
- EPA (1)
- Echtzeitsysteme (1)
- Elektronische Patientenakte (1)
- Ereignisse (1)
- Erfüllbarkeitsanalyse (1)
- Erkennen von Meta-Daten (1)
- Evolution (1)
- Evolution in MDE (1)
- Extract-Transform-Load (ETL) (1)
- FRP (1)
- Fallstudie (1)
- Feedback Loops (1)
- Fehlersuche (1)
- Formale Verifikation (1)
- Functional Lenses (1)
- Geschäftsprozesse (1)
- Geschäftsprozessmanagement (1)
- Gesetze (1)
- Graph-Constraints (1)
- Graph-basierte Suche (1)
- Graphbedingungen (1)
- Graphdatenbanken (1)
- Graphtransformation (1)
- HENSHIN (1)
- Hasso-Plattner-Institute (1)
- Hauptspeicherdatenbank (1)
- Homomorphe Verschlüsselung (1)
- In-Memory Database (1)
- In-Memory Datenbank (1)
- In-Memory-Datenbank (1)
- Individuen (1)
- Infinite State (1)
- Information Extraction (1)
- Information Systems (1)
- Informationsextraktion (1)
- Informationssysteme (1)
- Inkrementelle Graphmustersuche (1)
- Innovation (1)
- Innovationsmanagement (1)
- Innovationsmethode (1)
- Invariant-Checking (1)
- Invarianten (1)
- Invariants (1)
- JCop (1)
- Java (1)
- Kollaborationen (1)
- Laufzeitanalyse (1)
- Laufzeitmodelle (1)
- Leadership (1)
- Link Discovery (1)
- Link-Entdeckung (1)
- Linked Data (1)
- Linked Open Data (1)
- Live-Programmierung (1)
- Lively Kernel (1)
- MDE Ansatz (1)
- MDE settings (1)
- MOOCs (1)
- Management (1)
- Megamodell (1)
- Megamodels (1)
- Mehrkernsysteme (1)
- Metadata Discovery (1)
- Metadatenentdeckung (1)
- Metadatenqualität (1)
- Middleware (1)
- Mobile Application Development (1)
- Model Execution (1)
- Modeling Languages (1)
- Modell-getriebene Softwareentwicklung (1)
- Modellerzeugung (1)
- Modellgetriebene Softwareentwicklung (1)
- Modellierungssprachen (1)
- Modelltransformationen (1)
- Models at Runtime (1)
- Morphic (1)
- Multi-Instanzen (1)
- Multicore architectures (1)
- Muster (1)
- Musterabgleich (1)
- Nested Graph Conditions (1)
- Newspeak (1)
- Object Constraint Programming (1)
- Object-Oriented Programming (1)
- Objekt-Constraint Programmierung (1)
- Objekt-Orientiertes Programmieren (1)
- Objekt-orientiertes Programmieren mit Constraints (1)
- Objektlebenszyklus-Synchronisation (1)
- Online Course (1)
- Online-Learning (1)
- Online-Lernen (1)
- Onlinekurs (1)
- Organisationsveränderung (1)
- PRISM Modell-Checker (1)
- PRISM model checker (1)
- PTCTL (1)
- Pattern Matching (1)
- Patterns (1)
- Petri net Mapping (1)
- Petri net mapping (1)
- Petrinetz (1)
- Process Enactment (1)
- Process Mining (1)
- Programmierung (1)
- Programming Languages (1)
- Propagation von Aktivitätsinstanzzuständen (1)
- Prozessausführung (1)
- Prozesserhebung (1)
- Prozessinstanz (1)
- Prozessoren (1)
- Quantitative Analysen (1)
- Research Projects (1)
- Runtime Binding (1)
- SQL (1)
- STG decomposition (1)
- STG-Dekomposition (1)
- Sammlungsdatentypen (1)
- Schemaentdeckung (1)
- Schlüsselentdeckung (1)
- Self-Adaptive Software (1)
- Sequenzen von s/t-Pattern (1)
- Service-Oriented Architecture (1)
- Service-Orientierte Architekturen (1)
- Service-orientierte Systme (1)
- Signalflankengraph (SFG oder STG) (1)
- Simulation (1)
- Smalltalk (1)
- SoaML (1)
- Softwarearchitektur (1)
- Softwareproduktlinien (1)
- Softwaretests (1)
- Sozialen Medien (1)
- Speicheroptimierungen (1)
- Sprachspezifikation (1)
- Standards (1)
- Studie (1)
- Synchronisation (1)
- System of Systems (1)
- Systemsoftware (1)
- Tableaumethode (1)
- Tele-Lab (1)
- Tele-Teaching (1)
- Threshold Cryptography (1)
- Transformationsebene (1)
- Transformationssequenzen (1)
- Travis CI (1)
- Triple Graph Grammar (1)
- Triple Graph Grammars (1)
- Triple-Graph-Grammatiken (1)
- Unbegrenzter Zustandsraum (1)
- Unveränderlichkeit (1)
- Verbindungsnetzwerke (1)
- Verhaltensabstraktion (1)
- Verhaltensbewahrung (1)
- Verhaltensverfeinerung (1)
- Verhaltensäquivalenz (1)
- Verification (1)
- Verteilungsalgorithmen (1)
- Verteilungsalgorithmus (1)
- Verzögerungs-Verbreitung (1)
- Virtual machines (1)
- Visualisierung (1)
- Wartung von Graphdatenbanksichten (1)
- Web applications (1)
- Web-Anwendungen (1)
- Wikipedia (1)
- Zugriffskontrolle (1)
- access control (1)
- activity instance state propagation (1)
- adaptive Systeme (1)
- adaptive systems (1)
- adoption (1)
- apriori (1)
- architecture (1)
- ausführbare Semantiken (1)
- batch processing (1)
- behavior preservation (1)
- behavioral abstraction (1)
- behavioral equivalenc (1)
- behavioral refinement (1)
- beschreibende Feldstudie (1)
- big data services (1)
- bisimulation (1)
- bitcoin (1)
- business process management (1)
- business process model abstraction (1)
- business processes (1)
- case study (1)
- change management (1)
- cloud (1)
- cloud security (1)
- cloud storage (1)
- collaboration (1)
- collection types (1)
- confidentiality (1)
- conformance checking (1)
- continuous integration (1)
- continuous testing (1)
- control resynthesis (1)
- controlled experiment (1)
- cyber-physical systems (1)
- data center management (1)
- data flow correctness (1)
- data integration (1)
- deadline propagation (1)
- debugging (1)
- delay propagation (1)
- dependable computing (1)
- dependencies (1)
- design thinking (1)
- differential privacy (1)
- diffusion (1)
- direct manipulation (1)
- direkte Manipulation (1)
- discrimination networks (1)
- distribution algorithm (1)
- dynamic typing (1)
- dynamic programming languages (1)
- dynamische Programmiersprachen (1)
- dynamische Sprachen (1)
- eindeutig (1)
- electronic health record (1)
- erfahrbare Medien (1)
- events (1)
- evolution in MDE (1)
- executable semantics (1)
- feedback loops (1)
- fehlende Daten (1)
- formal verification (1)
- formal verification methods (1)
- formale Verifikation (1)
- functional dependency (1)
- functional lenses (1)
- functional programming (1)
- funktionale Abhängigkeit (1)
- funktionale Programmierung (1)
- future SOC lab (1)
- graph databases (1)
- graph queries (1)
- graph transformation systems (1)
- graph transformations (1)
- homomorphic encryption (1)
- hybrid graph-transformation-systems (1)
- hybride Graph-Transformations-Systeme (1)
- immutable values (1)
- in-memory database (1)
- individuals (1)
- inductive invariant checking (1)
- induktives Invariant Checking (1)
- inkrementelles Graph Pattern Matching (1)
- innovation (1)
- innovation capabilities (1)
- innovation management (1)
- interconnect (1)
- invariant checking (1)
- k-Induktion (1)
- k-induction (1)
- k-inductive invariant checking (1)
- k-inductive invariants (1)
- k-induktive Invarianten (1)
- k-induktives Invariant-Checking (1)
- key discovery (1)
- kontinuierliche Integration (1)
- kontinuierliches Testen (1)
- kontrolliertes Experiment (1)
- language specification (1)
- law (1)
- leadership (1)
- live programming (1)
- location-based (1)
- management (1)
- many-core (1)
- mehrdimensionale Belangtrennung (1)
- memory optimization (1)
- metadata discovery (1)
- metadata quality (1)
- middleware (1)
- missing data (1)
- model generation (1)
- model transformation (1)
- model-driven engineering (1)
- monitoring (1)
- morphic (1)
- multi-core (1)
- multi-dimensional separation of concerns (1)
- multi-instances (1)
- nested application conditions (1)
- object life cycle synchronization (1)
- object-constraint programming (1)
- openHPI (1)
- operating systems (1)
- organizational change (1)
- orts-basiert (1)
- parallel computing (1)
- paralleles Rechnen (1)
- partial application conditions (1)
- partielle Anwendungsbedingungen (1)
- periodic tasks (1)
- periodische Aufgaben (1)
- petri net (1)
- probabilistic timed automata (1)
- probabilistische zeitbehaftete Automaten (1)
- process elicitation (1)
- process instance (1)
- process mining (1)
- processor hardware (1)
- profiling (1)
- programming (1)
- quantitative analysis (1)
- reactive (1)
- reaktive Programmierung (1)
- real-time systems (1)
- relational model transformation (1)
- relationale Modelltransformationen (1)
- resource optimization (1)
- runtime adaptations (1)
- s/t-pattern sequences (1)
- satisfiabilitiy solving (1)
- schema discovery (1)
- semantics preservation (1)
- service-oriented systems (1)
- signal transition graph (1)
- simulation (1)
- smalltalk (1)
- software architecture (1)
- software product lines (1)
- software tests (1)
- speed independent (1)
- standards (1)
- static analysis (1)
- statische Analyse (1)
- stochastic Petri nets (1)
- stochastische Petri Netze (1)
- study (1)
- synchronization (1)
- system of systems (1)
- systems software (1)
- t.BPM (1)
- tableau method (1)
- tangible media (1)
- tele-TASK (1)
- threshold cryptography (1)
- transformation level (1)
- transformation sequences (1)
- triple graph grammars (1)
- typed graph transformation systems (1)
- unique (1)
- verschachtelte Anwednungsbedingungen (1)
- verschachtelte Graphbedingungen (1)
- view maintenance (1)
- visualization (1)
- web-applications (1)
- zuverlässige Datenverarbeitung (1)
- zuverlässigen Datenverarbeitung (1)
- Überwachung (1)
Institut
- Hasso-Plattner-Institut für Digital Engineering gGmbH (89) (entfernen)
Unique column combinations of a relational database table are sets of columns that contain only unique values. Discovering such combinations is a fundamental research problem and has many different data management and knowledge discovery applications. Existing discovery algorithms are either brute force or have a high memory load and can thus be applied only to small datasets or samples. In this paper, the wellknown GORDIAN algorithm and "Apriori-based" algorithms are compared and analyzed for further optimization. We greatly improve the Apriori algorithms through efficient candidate generation and statistics-based pruning methods. A hybrid solution HCAGORDIAN combines the advantages of GORDIAN and our new algorithm HCA, and it significantly outperforms all previous work in many situations.
Extract-Transform-Load (ETL) tools are used for the creation, maintenance, and evolution of data warehouses, data marts, and operational data stores. ETL workflows populate those systems with data from various data sources by specifying and executing a DAG of transformations. Over time, hundreds of individual workflows evolve as new sources and new requirements are integrated into the system. The maintenance and evolution of large-scale ETL systems requires much time and manual effort. A key problem is to understand the meaning of unfamiliar attribute labels in source and target databases and ETL transformations. Hard-to-understand attribute labels lead to frustration and time spent to develop and understand ETL workflows. We present a schema decryption technique to support ETL developers in understanding cryptic schemata of sources, targets, and ETL transformations. For a given ETL system, our recommender-like approach leverages the large number of mapped attribute labels in existing ETL workflows to produce good and meaningful decryptions. In this way we are able to decrypt attribute labels consisting of a number of unfamiliar few-letter abbreviations, such as UNP_PEN_INT, which we can decrypt to UNPAID_PENALTY_INTEREST. We evaluate our schema decryption approach on three real-world repositories of ETL workflows and show that our approach is able to suggest high-quality decryptions for cryptic attribute labels in a given schema.
Program behavior that relies on contextual information, such as physical location or network accessibility, is common in today's applications, yet its representation is not sufficiently supported by programming languages. With context-oriented programming (COP), such context-dependent behavioral variations can be explicitly modularized and dynamically activated. In general, COP could be used to manage any context-specific behavior. However, its contemporary realizations limit the control of dynamic adaptation. This, in turn, limits the interaction of COP's adaptation mechanisms with widely used architectures, such as event-based, mobile, and distributed programming. The JCop programming language extends Java with language constructs for context-oriented programming and additionally provides a domain-specific aspect language for declarative control over runtime adaptations. As a result, these redesigned implementations are more concise and better modularized than their counterparts using plain COP. JCop's main features have been described in our previous publications. However, a complete language specification has not been presented so far. This report presents the entire JCop language including the syntax and semantics of its new language constructs.
Data dependencies, or integrity constraints, are used to improve the quality of a database schema, to optimize queries, and to ensure consistency in a database. In the last years conditional dependencies have been introduced to analyze and improve data quality. In short, a conditional dependency is a dependency with a limited scope defined by conditions over one or more attributes. Only the matching part of the instance must adhere to the dependency. In this paper we focus on conditional inclusion dependencies (CINDs). We generalize the definition of CINDs, distinguishing covering and completeness conditions. We present a new use case for such CINDs showing their value for solving complex data quality tasks. Further, we define quality measures for conditions inspired by precision and recall. We propose efficient algorithms that identify covering and completeness conditions conforming to given quality thresholds. Our algorithms choose not only the condition values but also the condition attributes automatically. Finally, we show that our approach efficiently provides meaningful and helpful results for our use case.
Data obtained from foreign data sources often come with only superficial structural information, such as relation names and attribute names. Other types of metadata that are important for effective integration and meaningful querying of such data sets are missing. In particular, relationships among attributes, such as foreign keys, are crucial metadata for understanding the structure of an unknown database. The discovery of such relationships is difficult, because in principle for each pair of attributes in the database each pair of data values must be compared. A precondition for a foreign key is an inclusion dependency (IND) between the key and the foreign key attributes. We present with Spider an algorithm that efficiently finds all INDs in a given relational database. It leverages the sorting facilities of DBMS but performs the actual comparisons outside of the database to save computation. Spider analyzes very large databases up to an order of magnitude faster than previous approaches. We also evaluate in detail the effectiveness of several heuristics to reduce the number of necessary comparisons. Furthermore, we generalize Spider to find composite INDs covering multiple attributes, and partial INDs, which are true INDs for all but a certain number of values. This last type is particularly relevant when integrating dirty data as is often the case in the life sciences domain - our driving motivation.
Cyber-physical systems achieve sophisticated system behavior exploring the tight interconnection of physical coupling present in classical engineering systems and information technology based coupling. A particular challenging case are systems where these cyber-physical systems are formed ad hoc according to the specific local topology, the available networking capabilities, and the goals and constraints of the subsystems captured by the information processing part. In this paper we present a formalism that permits to model the sketched class of cyber-physical systems. The ad hoc formation of tightly coupled subsystems of arbitrary size are specified using a UML-based graph transformation system approach. Differential equations are employed to define the resulting tightly coupled behavior. Together, both form hybrid graph transformation systems where the graph transformation rules define the discrete steps where the topology or modes may change, while the differential equations capture the continuous behavior in between such discrete changes. In addition, we demonstrate that automated analysis techniques known for timed graph transformation systems for inductive invariants can be extended to also cover the hybrid case for an expressive case of hybrid models where the formed tightly coupled subsystems are restricted to smaller local networks.
Service-oriented modeling employs collaborations to capture the coordination of multiple roles in form of service contracts. In case of dynamic collaborations the roles may join and leave the collaboration at runtime and therefore complex structural dynamics can result, which makes it very hard to ensure their correct and safe operation. We present in this paper our approach for modeling and verifying such dynamic collaborations. Modeling is supported using a well-defined subset of UML class diagrams, behavioral rules for the structural dynamics, and UML state machines for the role behavior. To be also able to verify the resulting service-oriented systems, we extended our former results for the automated verification of systems with structural dynamics [7, 8] and developed a compositional reasoning scheme, which enables the reuse of verification results. We outline our approach using the example of autonomous vehicles that use such dynamic collaborations via ad-hoc networking to coordinate and optimize their joint behavior.
Graph queries have lately gained increased interest due to application areas such as social networks, biological networks, or model queries. For the relational database case the relational algebra and generalized discrimination networks have been studied to find appropriate decompositions into subqueries and ordering of these subqueries for query evaluation or incremental updates of query results. For graph database queries however there is no formal underpinning yet that allows us to find such suitable operationalizations. Consequently, we suggest a simple operational concept for the decomposition of arbitrary complex queries into simpler subqueries and the ordering of these subqueries in form of generalized discrimination networks for graph queries inspired by the relational case. The approach employs graph transformation rules for the nodes of the network and thus we can employ the underlying theory. We further show that the proposed generalized discrimination networks have the same expressive power as nested graph conditions.