Refine
Year of publication
Document Type
- Article (181)
- Monograph/Edited Volume (89)
- Doctoral Thesis (42)
- Other (29)
- Conference Proceeding (11)
- Preprint (4)
- Postprint (2)
- Review (2)
- Part of a Book (1)
Language
- English (361) (remove)
Keywords
- Cloud Computing (7)
- Hasso-Plattner-Institut (7)
- Datenintegration (6)
- Forschungskolleg (6)
- Hasso Plattner Institute (6)
- Klausurtagung (6)
- Modellierung (6)
- Service-oriented Systems Engineering (6)
- cloud computing (6)
- Forschungsprojekte (5)
- Future SOC Lab (5)
- In-Memory Technologie (5)
- Multicore Architekturen (5)
- data profiling (5)
- machine learning (5)
- BPMN (4)
- COVID-19 (4)
- Geschäftsprozessmanagement (4)
- Model-Driven Engineering (4)
- Research School (4)
- Verifikation (4)
- Visualization (4)
- business process management (4)
- data integration (4)
- graph transformation (4)
- middleware (4)
- performance (4)
- process mining (4)
- virtual machines (4)
- 3D point clouds (3)
- Betriebssysteme (3)
- Data Integration (3)
- Graphtransformationen (3)
- Model Synchronisation (3)
- Model Transformation (3)
- Modeling (3)
- Ph.D. Retreat (3)
- Ph.D. retreat (3)
- Privacy (3)
- Process Mining (3)
- Prozessmodellierung (3)
- SQL (3)
- Security (3)
- Tripel-Graph-Grammatik (3)
- Virtualisierung (3)
- Virtuelle Maschinen (3)
- cartographic design (3)
- design thinking (3)
- digital health (3)
- duplicate detection (3)
- heuristics (3)
- incremental graph pattern matching (3)
- model transformation (3)
- multicore architectures (3)
- operating systems (3)
- prediction (3)
- privacy (3)
- programming (3)
- research projects (3)
- run time analysis (3)
- security (3)
- self-healing (3)
- service-oriented systems engineering (3)
- similarity measures (3)
- theory (3)
- verification (3)
- 3D city models (2)
- 3D geovisualization (2)
- AUTOSAR (2)
- Abstraktion (2)
- Algorithms (2)
- Aspektorientierte Softwareentwicklung (2)
- Assoziationsregeln (2)
- Asynchrone Schaltung (2)
- BIM (2)
- Bayesian networks (2)
- Big Data (2)
- Bitcoin (2)
- Blockchain (2)
- CSC (2)
- CSCW (2)
- Case management (2)
- Classification (2)
- Cloud-Sicherheit (2)
- Cloud-Speicher (2)
- Competition (2)
- Compliance checking (2)
- Cyber-Physical Systems (2)
- Data Profiling (2)
- Data integration (2)
- Data profiling (2)
- Design Thinking (2)
- Discrimination Networks (2)
- EHR (2)
- Event processing (2)
- Evolution (2)
- Functional dependencies (2)
- Graphtransformationssysteme (2)
- HCI (2)
- In-Memory technology (2)
- Java (2)
- JavaScript (2)
- Kollaborationen (2)
- Laufzeitmodelle (2)
- Link-Entdeckung (2)
- Lively Kernel (2)
- Machine (2)
- Machine learning (2)
- Megamodell (2)
- Metadata Discovery (2)
- Middleware (2)
- Model Synchronization (2)
- Modell (2)
- Nested graph conditions (2)
- Performance (2)
- Point-based rendering (2)
- Process Modeling (2)
- Process modeling (2)
- RDF (2)
- Ressourcenoptimierung (2)
- Runtime analysis (2)
- SPARQL (2)
- STG decomposition (2)
- STG-Dekomposition (2)
- Service-Orientierte Architekturen (2)
- Sicherheit (2)
- Smalltalk (2)
- Structuring (2)
- SysML (2)
- Systemsoftware (2)
- Virtualization (2)
- Visualisierung (2)
- abstraction (2)
- adaptive Systeme (2)
- adaptive systems (2)
- analysis (2)
- big data services (2)
- clinical (2)
- cloud security (2)
- collaboration (2)
- complexity (2)
- confidentiality (2)
- data (2)
- data matching (2)
- data quality (2)
- data wrangling (2)
- databases (2)
- debugging (2)
- dependency discovery (2)
- dialysis (2)
- digital identity (2)
- discrimination networks (2)
- distributed systems (2)
- dynamic (2)
- dynamic programming (2)
- electronic health record (2)
- entity resolution (2)
- evaluation (2)
- fabrication (2)
- feedback loops (2)
- functional dependency (2)
- genetic programming (2)
- graph constraints (2)
- image-based representation (2)
- in-memory technology (2)
- law (2)
- missing data (2)
- model (2)
- model-driven engineering (2)
- modeling (2)
- modellgetriebene Entwicklung (2)
- multi-core (2)
- nested application conditions (2)
- nested graph conditions (2)
- non-photorealistic rendering (2)
- real-time rendering (2)
- record linkage (2)
- requirements engineering (2)
- research school (2)
- risk aversion (2)
- scalability (2)
- schema discovery (2)
- self-sovereign identity (2)
- service-oriented architecture (2)
- service-oriented systems (2)
- simulation (2)
- software engineering (2)
- software reference architecture (2)
- spatial data infrastructure (2)
- standardization (2)
- stochastic Petri nets (2)
- stochastische Petri Netze (2)
- systems of systems (2)
- systems software (2)
- testing (2)
- virtualization (2)
- virtuelle Maschinen (2)
- visualization (2)
- "Big Data"-Dienste (1)
- 2.5D Treemaps (1)
- 3D (1)
- 3D Computer Grafik (1)
- 3D Computer Graphics (1)
- 3D Drucken (1)
- 3D Point Clouds (1)
- 3D Point clouds (1)
- 3D Semiotik (1)
- 3D Visualisierung (1)
- 3D geovirtual environments (1)
- 3D information visualization (1)
- 3D printing (1)
- 3D semiotic model (1)
- 3D semiotics (1)
- 3D visualization (1)
- AKI (1)
- APX-hardness (1)
- Aaron Wildavsky (1)
- Abhängigkeiten (1)
- Abstraktion von Geschäftsprozessmodellen (1)
- Accounting (1)
- Actor (1)
- Actor model (1)
- Address matching (1)
- Algorithmen (1)
- Analyse (1)
- Anfragepaare (1)
- Anisotroper Kuwahara Filter (1)
- Anomalien (1)
- Anomaly detection (1)
- Anti-patterns (1)
- Application (1)
- Approximation algorithms (1)
- Apriori (1)
- Architektur (1)
- Artificial neural networks (1)
- Aspect-oriented Programming (1)
- Association Rule Mining (1)
- Association rule mining (1)
- Asynchronous circuit (1)
- Attention span (1)
- Attribut-Merge-Prozess (1)
- Attribute Merge Process (1)
- Attribute aggregation (1)
- Attributed graph transformation (1)
- Attributed graphs (1)
- Ausführung von Modellen (1)
- Ausführungsgeschichte (1)
- Authentication (1)
- B2B process integration (1)
- BPM (1)
- BPMN-Q (1)
- Batchprozesse (1)
- Bayes'sche Netze (1)
- Bayessche Netze (1)
- Bedingte Inklusionsabhängigkeiten (1)
- Behavior (1)
- Behavior equivalence (1)
- Behavioral querying (1)
- Behaviour Analysis (1)
- Behavioural Abstraction (1)
- Behavioural analysis (1)
- Benchmarking (1)
- Berührungseingaben (1)
- Beschränkungen und Abhängigkeiten (1)
- Biclustering (1)
- Bidirectional order dependencies (1)
- Bildverarbeitung (1)
- Billing (1)
- Biomedicine (1)
- Bisimulation (1)
- Blockchains (1)
- Bluetooth (1)
- Body sensor networks (1)
- Building Information Models (1)
- Business Process Modeling Notation (1)
- Business Process Models (1)
- Business process diagram (1)
- Business process management (1)
- Business process model (1)
- Business process modeling (1)
- Business processes (1)
- CCS Concepts (1)
- CEP (1)
- Cake cutting (1)
- Carrera Digital D132 (1)
- Causal Behavioural Profiles (1)
- Causality (1)
- Change Data Capture (1)
- Change Management (1)
- Change propagation (1)
- Charging (1)
- Choreographies (1)
- Chronic heart failure (1)
- Climate change (1)
- Close-Up (1)
- Cloud (1)
- Cloud Datenzentren (1)
- Cloud Native Applications (1)
- Cloud Storage Broker (1)
- Cloud access control and resource management (1)
- Cloud computing (1)
- Cloud-Security (1)
- CoExist (1)
- Coccinelle (1)
- Communication systems (1)
- Commute pattern (1)
- Commute process (1)
- Complexity theory (1)
- Compliance (1)
- Compliance measurement (1)
- Composition (1)
- Computational modeling (1)
- Computational photography (1)
- Computer (1)
- Computer crime (1)
- Concurrency (1)
- Conditional Inclusion Dependency (1)
- Conformance Überprüfung (1)
- Consistency (1)
- Consistency perception (1)
- Constraints (1)
- Context-oriented Programming (1)
- Context-oriented programming (1)
- ContextJS (1)
- Contracts (1)
- Controller-Resynthese (1)
- Convolution (1)
- Coordinated and Multiple Views (1)
- CorMID (1)
- Correlation (1)
- Critical pairs (1)
- Crowd-Resourcing (1)
- Cryptography (1)
- Cultural theory (1)
- Currencies (1)
- Cyber-Physical-Systeme (1)
- Cyber-physical-systems (1)
- Data (1)
- Data Dependency (1)
- Data Mining (1)
- Data Modeling (1)
- Data Quality (1)
- Data Warehouse (1)
- Data dependencies (1)
- Data exchange (1)
- Data mining (1)
- Data modeling (1)
- Data models (1)
- Data-centric (1)
- Database Cost Model (1)
- Databases (1)
- Daten (1)
- Datenabhängigkeiten (1)
- Datenabhängigkeiten-Entdeckung (1)
- Datenanalyse (1)
- Datenbank-Kostenmodell (1)
- Datenbanken (1)
- Datenextraktion (1)
- Datenflusskorrektheit (1)
- Datenkorrektheit (1)
- Datenmodellierung (1)
- Datenobjekte (1)
- Datenqualität (1)
- Datenreinigung (1)
- Datenvertraulichkeit (1)
- Datenzustände (1)
- Deadline-Verbreitung (1)
- Deep learning (1)
- Delphi study (1)
- Delta preservation (1)
- Dependency discovery (1)
- Design (1)
- Detail plus Overview (1)
- Deurema Modellierungssprache (1)
- Differential Privacy (1)
- Differenz von Gauss Filtern (1)
- Digitale Whiteboards (1)
- Disambiguierung (1)
- Distributed (1)
- Distributed 3D geovisualization (1)
- Distributed computing (1)
- Distributed debugging (1)
- Distributed programming (1)
- DoS (1)
- Duplicate Detection (1)
- Duplikaterkennung (1)
- Duration prediction (1)
- Dynamic Data (1)
- Dynamic Data Structures (1)
- Dynamic Pricing (1)
- Dynamic Type System (1)
- Dynamic adaptation (1)
- Dynamic analysis (1)
- Dynamic pricing (1)
- Dynamic pricing and advertising (1)
- Dynamische Typ Systeme (1)
- E-Learning (1)
- E-commerce (1)
- EPA (1)
- Echtzeit (1)
- Echtzeitsysteme (1)
- Ecosystems (1)
- Eingabegenauigkeit (1)
- Electrocardiography (1)
- Electronic prescription (1)
- Elektronische Patientenakte (1)
- Energieeffizienz (1)
- Energy-aware (1)
- Entity resolution (1)
- Entwicklungswerkzeuge (1)
- Entwurfsmuster für SOA-Sicherheit (1)
- Ereignisabstraktion (1)
- Ereignisse (1)
- Erfahrungsbericht (1)
- Erfüllbarkeitsanalyse (1)
- Erkennen von Meta-Daten (1)
- Estimation-of-distribution algorithm (1)
- Event normalization (1)
- Events (1)
- Evolution in MDE (1)
- Evolutionary algorithms (1)
- Evolutionary computation (1)
- Exception handling (1)
- Exclusiveness (1)
- Experimentation (1)
- Extract-Transform-Load (ETL) (1)
- Eye-tracking (1)
- FMC-QE (1)
- FRP (1)
- Facial mimicry (1)
- Fallstudie (1)
- Feature extraction (1)
- Feature selection (1)
- Federated learning (1)
- Feedback Loop Modellierung (1)
- Feedback Loops (1)
- Feedback heuristics (1)
- Fehlende Daten (1)
- Fehlerbeseitigung (1)
- Fehlersuche (1)
- Finite horizon (1)
- First hitting time (1)
- Fitness level method (1)
- Fitness-distance correlation (1)
- Flexible Resource Manager (1)
- Flexible processes; (1)
- Flussgesteuerter Bilateraler Filter (1)
- Focus+Context Visualization (1)
- Fokus-&-Kontext Visualisierung (1)
- Foreign Keys (1)
- Foreign Keys Discovery (1)
- Formal Methods (1)
- Formale Verifikation (1)
- Functional Lenses (1)
- GPU (1)
- GPU-based Real-time Rendering (1)
- Gamification (1)
- Gator Netzwerk (1)
- Gator networks (1)
- Gebäudemodelle (1)
- Geländemodelle (1)
- Gender Inequality (1)
- Gene expression (1)
- General Earth and Planetary Sciences (1)
- Generalisierung (1)
- Geodaten (1)
- Geography, Planning and Development (1)
- Geometry Draping (1)
- Geovisualization (1)
- Geschäftsprozesse (1)
- Geschäftsprozessmodelle (1)
- Gesetze (1)
- Graph databases (1)
- Graph homomorphisms (1)
- Graph queries (1)
- Graph repair (1)
- Graph rewriting (1)
- Graph transformation (1)
- Graph-Constraints (1)
- Graph-basierte Suche (1)
- Graphbedingungen (1)
- Graphdatenbanken (1)
- Graphtransformation (1)
- HENSHIN (1)
- HITS (1)
- HMM (1)
- Haptics (1)
- Hasso-Plattner-Institute (1)
- Hauptspeicher Technologie (1)
- Hauptspeicherdatenbank (1)
- Herodotos (1)
- HiGHmed (1)
- Hidden Markov models (1)
- History of pattern occurrences (1)
- Homomorphe Verschlüsselung (1)
- Hospitalisation (1)
- Human behaviour (1)
- IDS (1)
- IDS management (1)
- IFC (1)
- IMD (1)
- IMU (1)
- IOPS (1)
- IT-Security (1)
- IT-Sicherheit (1)
- Identity management systems (1)
- Image (1)
- Image filtering (1)
- Image-based rendering (1)
- In-Memory Database (1)
- In-Memory Datenbank (1)
- In-Memory-Datenbank (1)
- In-memory (1)
- Inclusion Dependency (1)
- Inclusion Dependency Discovery (1)
- Inclusion dependencies (1)
- Incremental Discovery (1)
- Incrementally Inclusion Dependencies Discovery (1)
- Index (1)
- Index Structures (1)
- Indexstrukturen (1)
- Individuen (1)
- Indoor Models (1)
- Industries (1)
- Industry 4.0 (1)
- Industry Foundation Classes (1)
- Infinite State (1)
- Information Extraction (1)
- Information Systems (1)
- Information Visualization (1)
- Informationsextraktion (1)
- Informationssysteme (1)
- Informationsvorhaltung (1)
- Initial conflicts (1)
- Inklusionsabhängigkeit (1)
- Inklusionsabhängigkeiten (1)
- Inklusionsabhängigkeiten Entdeckung (1)
- Inkrementelle Graphmustersuche (1)
- Innovation (1)
- Innovationsmanagement (1)
- Innovationsmethode (1)
- Integrity Verification (1)
- Interaction (1)
- Interaction modeling (1)
- Interactive Rendering (1)
- Interactive Visualization (1)
- Interaktionsmodel (1)
- Interaktives Rendering (1)
- Internet applications (1)
- Internet of Things (1)
- Internetanwendungen (1)
- Interviews (1)
- Introspection (1)
- Intrusion detection (1)
- Invariant-Checking (1)
- Invarianten (1)
- Invariants (1)
- Inventory holding costs (1)
- Inventory systems (1)
- JCop (1)
- JIT compilers (1)
- Kartografisches Design (1)
- Knowledge bases (1)
- Knowledge-intensive processes (1)
- Komplexität (1)
- Komplexitätsbewältigung (1)
- Komposition (1)
- Kontext (1)
- LEGO Mindstorms EV3 (1)
- LIDAR (1)
- LOD (1)
- LSTM (1)
- Label analysis (1)
- Lakes (1)
- Landmarken (1)
- Languages (1)
- Languages Model-driven engineering (1)
- Laser Cutten (1)
- Laser cutting (1)
- Laufzeitanalyse (1)
- Laufzeitverhalten (1)
- Leadership (1)
- Learning (1)
- Learning behavior (1)
- Least privilege principle (1)
- Lecture video recording (1)
- Leistungsfähigkeit (1)
- Leistungsvorhersage (1)
- Level of abstraction (1)
- Licenses (1)
- Link Discovery (1)
- Linked Data (1)
- Linked Open Data (1)
- Live forensics (1)
- Live-Programmierung (1)
- Location-based services (1)
- Log conformance (1)
- Logiksynthese (1)
- M-adhesive categories (1)
- M-adhesive transformation systems (1)
- MDE Ansatz (1)
- MDE settings (1)
- MOOC (1)
- MOOCs (1)
- Management (1)
- Markov decision process (1)
- Markov decision process; (1)
- Markov model (1)
- Mary Douglas (1)
- Massive Open Online Courses (1)
- Matroids (1)
- Measurement (1)
- Megamodel (1)
- Megamodels (1)
- Mehrkernsysteme (1)
- Memory management (1)
- Metadaten Entdeckung (1)
- Metadatenentdeckung (1)
- Metadatenqualität (1)
- Metamaterials (1)
- Mind2 (1)
- Mixed workload (1)
- Mobile Application Development (1)
- Mobile device (1)
- Mobile sensing (1)
- Mobilgeräte (1)
- Model Consistency (1)
- Model Execution (1)
- Model Management (1)
- Model equivalence (1)
- Model generation (1)
- Model refinement (1)
- Model repair (1)
- Model synchronisation (1)
- Model transformation (1)
- Model verification (1)
- Model-driven (1)
- Modeling Languages (1)
- Modell Management (1)
- Modell-driven Security (1)
- Modell-getriebene Sicherheit (1)
- Modell-getriebene Softwareentwicklung (1)
- Modellerzeugung (1)
- Modellgetrieben (1)
- Modellgetriebene Entwicklung (1)
- Modellgetriebene Softwareentwicklung (1)
- Modellierungssprachen (1)
- Modellkonsistenz (1)
- Modelltransformation (1)
- Modelltransformationen (1)
- Models at Runtime (1)
- Modular decomposition (1)
- Monitoring language (1)
- Morphic (1)
- Multi-Instanzen (1)
- Multi-objective optimization (1)
- Multi-perspective Views (1)
- Multicore architectures (1)
- Multimedia retrieval (1)
- Multiscale modeling (1)
- Muster (1)
- Musterabgleich (1)
- Mustererkennung (1)
- Mutation operators (1)
- N-of-1 trial (1)
- NP-completeness (1)
- Natural language (1)
- Natural language processing (1)
- Nested Graph Conditions (1)
- Network graph (1)
- Network monitoring (1)
- Network topology (1)
- Neural Networks (1)
- Newspeak (1)
- Nicht-photorealistisches Rendering (1)
- Nichtfotorealistische Bildsynthese (1)
- Nigeria (1)
- Nutzungsinteresse (1)
- O (1)
- OLTP (1)
- Object Constraint Programming (1)
- Object Versioning (1)
- Object-Oriented Programming (1)
- Objekt-Constraint Programmierung (1)
- Objekt-Orientiertes Programmieren (1)
- Objekt-orientiertes Programmieren mit Constraints (1)
- Objektlebenszyklus-Synchronisation (1)
- Online Course (1)
- Online-Learning (1)
- Online-Lernen (1)
- Onlinekurs (1)
- Open implementations (1)
- Operational reporting (1)
- Optimal Control (1)
- Optimal stochastic and deterministic (1)
- Optionality (1)
- Order Relations (1)
- Order dependencies (1)
- Organisationsveränderung (1)
- Out-of-core (1)
- Outlier detection (1)
- Overview plus Detail (1)
- PPMI (parkinson's progression markers initiative) (1)
- PRISM Modell-Checker (1)
- PRISM model checker (1)
- PTCTL (1)
- Parallel programming (1)
- Parallelization (1)
- Pattern Matching (1)
- Patterns (1)
- Performance Prediction (1)
- Performance analysis (1)
- Personal fabrication (1)
- Petri net Mapping (1)
- Petri net mapping (1)
- Petri net unfolding (1)
- Petrinetz (1)
- Plugs (1)
- Prediction (1)
- Price Cycles (1)
- Price collusion (1)
- Privilege separation concept (1)
- Probabilistische Modelle (1)
- Process (1)
- Process Enactment (1)
- Process Modelling (1)
- Process Monitoring (1)
- Process choreographies (1)
- Process compliance (1)
- Process mining (1)
- Process model (1)
- Process model alignment (1)
- Process model consistency (1)
- Process model repositories (1)
- Process model search (1)
- Processing strategies (1)
- Programmierung (1)
- Programming (1)
- Programming Environments (1)
- Programming Languages (1)
- Propagation von Aktivitätsinstanzzuständen (1)
- Property paths (1)
- Protocols (1)
- Prozess (1)
- Prozess- und Datenintegration (1)
- Prozessarchitektur (1)
- Prozessausführung (1)
- Prozessautomatisierung (1)
- Prozesse (1)
- Prozesserhebung (1)
- Prozessinstanz (1)
- Prozessmodellsuche (1)
- Prozessoren (1)
- Prozessverfeinerung (1)
- Präsentation (1)
- QRS detection (1)
- Quality of service (1)
- Quantitative Analysen (1)
- Quantitative Modeling (1)
- Quantitative Modellierung (1)
- Query (1)
- Query execution (1)
- Query optimization (1)
- Question answering (1)
- Queuing Theory (1)
- R Shiny (1)
- R package (1)
- RT_PREEMT patch (1)
- RT_PREEMT-Patch (1)
- Racket (1)
- Random graphs (1)
- Reaction Time (1)
- Real-time Rendering (1)
- Record and refinement (1)
- Record and replay (1)
- Record linkage (1)
- Reinforcement learning (1)
- Relational data (1)
- Remote patient management (1)
- Research Projects (1)
- Resource description framework (1)
- Resource management (1)
- Response Strategies (1)
- Ressourcenmanagement (1)
- Rete Netzwerk (1)
- Rete networks (1)
- Risk control (1)
- Robust optimization (1)
- Role-based access control (1)
- Root cause analysis (1)
- Run time analysis (1)
- Runtime Binding (1)
- Runtime WCET Analysis (1)
- S-indd++ (1)
- SAP HANA (1)
- SCED (1)
- SOA Security Pattern (1)
- Safety Critical Systems (1)
- Sammlungsdatentypen (1)
- Satisfiability (1)
- Scalability (1)
- Schema-Entdeckung (1)
- Schemaentdeckung (1)
- Schlüsselentdeckung (1)
- Scientific Publication Indicators (1)
- Scope (1)
- Search Algorithms (1)
- Security Modelling (1)
- Security-as-a-Service (1)
- Self-Adaptive Software (1)
- Self-aware computing systems (1)
- Self-configuration (1)
- Semantics (1)
- Semantische Analyse (1)
- Sequential anomaly (1)
- Sequenzen von s/t-Pattern (1)
- Service composition (1)
- Service detection (1)
- Service orchestration (1)
- Service-Oriented (1)
- Service-Oriented Architecture (1)
- Service-oriented Architectures (1)
- Service-oriented computing (1)
- Service-orientierte Systeme (1)
- Service-orientierte Systme (1)
- Sexism (1)
- Sicherheitsmodellierung (1)
- Signal-to-noise ratio (1)
- Signalflankengraph (SFG oder STG) (1)
- Similarity Measures (1)
- Similarity Search (1)
- Simulation (1)
- Skalierbarkeit (1)
- SoaML (1)
- Social environment (1)
- Software (1)
- Software Engineering (1)
- Softwareanalyse (1)
- Softwarearchitektur (1)
- Softwareentwicklung (1)
- Softwareentwicklungsprozesse (1)
- Softwareproduktlinien (1)
- Softwaretechnik (1)
- Softwaretest (1)
- Softwaretests (1)
- Softwarevisualisierung (1)
- Softwarewartung (1)
- Sozialen Medien (1)
- Speicheroptimierungen (1)
- Sprachspezifikation (1)
- Squeak (1)
- Standards (1)
- Stilisierung (1)
- Stochastic Petri nets (1)
- Structural Decomposition (1)
- Structured modeling (1)
- Strukturierung (1)
- Studie (1)
- Submodular function (1)
- Submodular functions (1)
- Subset selection (1)
- Suchverfahren (1)
- Synchronisation (1)
- Synonyme (1)
- System architecture (1)
- System of Systems (1)
- Systeme von Systemen (1)
- Systems of Systems (1)
- TRIPOD (1)
- Tableau method (1)
- Tableaumethode (1)
- Task analysis (1)
- Tele-Lab (1)
- Tele-Teaching (1)
- Telemonitoring (1)
- Temporal Logic (1)
- Temporal orientation (1)
- Temporallogik (1)
- Terrain Visualization (1)
- Test-getriebene Fehlernavigation (1)
- Testen (1)
- Texturing (1)
- Theory (1)
- Three-dimensional displays (1)
- Threshold Cryptography (1)
- Time Augmented Petri Nets (1)
- Time series (1)
- Tool survey (1)
- Trace inclusion (1)
- Traceability (1)
- Tracking (1)
- Transformation (1)
- Transformation tool contest (1)
- Transformationsebene (1)
- Transformationssequenzen (1)
- Travis CI (1)
- Treemaps (1)
- Triple Graph Grammar (1)
- Triple Graph Grammars (1)
- Triple-Graph-Grammatiken (1)
- Two dimensional displays (1)
- Ubiquitous computing (1)
- Unbegrenzter Zustandsraum (1)
- Unified cloud model (1)
- Unique column combination (1)
- Unique column combinations (1)
- Unveränderlichkeit (1)
- Usage Interest (1)
- VIL (1)
- Verbindungsnetzwerke (1)
- Verhalten (1)
- Verhaltensabstraktion (1)
- Verhaltensanalyse (1)
- Verhaltensbewahrung (1)
- Verhaltensverfeinerung (1)
- Verhaltensäquivalenz (1)
- Verification (1)
- Verletzung Auflösung (1)
- Verletzung Erklärung (1)
- Vernetzte Daten (1)
- Versionierung (1)
- Verteiltes Arbeiten (1)
- Verteilungsalgorithmen (1)
- Verteilungsalgorithmus (1)
- Verwaltung von Rechenzentren (1)
- Verzögerungs-Verbreitung (1)
- Video OCR (1)
- Video classification (1)
- Video indexing (1)
- Videoanalyse (1)
- Videometadaten (1)
- View navigation (1)
- Violation Explanation (1)
- Violation Resolution (1)
- Virtual 3D city model (1)
- Virtual 3D scenes (1)
- Virtual Machine (1)
- Virtual camera control (1)
- Virtual machines (1)
- Visual modeling (1)
- Vocabulary (1)
- Vorhersage (1)
- Vulnerability Assessment (1)
- W[3]-completeness (1)
- Warteschlangentheorie (1)
- Wartung von Graphdatenbanksichten (1)
- Water Science and Technology (1)
- Web Sites (1)
- Web applications (1)
- Web browsers (1)
- Web navigational language (1)
- Web of Data (1)
- Web safeness (1)
- Web-Anwendungen (1)
- Web-based rendering (1)
- Webseite (1)
- Well-structuredness (1)
- Wikipedia (1)
- Wohlstrukturiertheit (1)
- Zeitbehaftete Petri Netze (1)
- Zugriffskontrolle (1)
- abdominal imaging (1)
- abundance (1)
- access control (1)
- accessibility (1)
- accounting (1)
- action recognition (1)
- activity instance state propagation (1)
- activity recognition (1)
- acute renal failure (1)
- adaptation rules (1)
- address normalization (1)
- address parsing (1)
- adolescent (1)
- adoption (1)
- affective interfaces (1)
- algorithm (EDA) (1)
- algorithmic (1)
- anisotropic Kuwahara filter (1)
- anniversary (1)
- anomalies (1)
- app (1)
- application (1)
- approximation (1)
- apriori (1)
- architecture (1)
- architecture recovery (1)
- architecture-based adaptation (1)
- architectures (1)
- argumentation research (1)
- artificial intelligence (1)
- aspect adapter (1)
- aspect oriented programming (1)
- aspect-oriented (1)
- aspects (1)
- aspectualization (1)
- association rule mining (1)
- asynchronous circuit (1)
- atmospheric pressure chemical ionization (1)
- attack graph (1)
- attacks (1)
- attention mechanism (1)
- attribute assurance (1)
- ausführbare Semantiken (1)
- authentication (1)
- authentication protocol (1)
- back pain (1)
- back-in-time (1)
- batch processing (1)
- battery (1)
- battery-depletion attack (1)
- behavior preservation (1)
- behavioral abstraction (1)
- behavioral equivalenc (1)
- behavioral refinement (1)
- behavioral specification (1)
- behaviour compatibility (1)
- behaviour equivalence (1)
- behavioural models (1)
- benchmark testing (1)
- beschreibende Feldstudie (1)
- bibliometrics (1)
- bidirectional shortest path (1)
- big data (1)
- billing (1)
- biomarker detection (1)
- biomechanics (1)
- bisimulation (1)
- bitcoin (1)
- blind (1)
- bloat control (1)
- bpm (1)
- bridge management systems (1)
- bug tracking (1)
- building information modeling (1)
- building models (1)
- business process architecture (1)
- business process model abstraction (1)
- business processes (1)
- cancer therapy (1)
- cartography-oriented visualization (1)
- case study (1)
- center dot Computing (1)
- change blindness (1)
- change management (1)
- changeability (1)
- charging (1)
- child (1)
- chronic dialysis (1)
- cleansing (1)
- clinical nephrology (1)
- cloud (1)
- cloud datacenter (1)
- cloud storage (1)
- cluster resource management (1)
- clustering (1)
- co-location (1)
- cognition (1)
- cognitive (1)
- coherence-enhancing filtering (1)
- cohort (1)
- collaborations (1)
- collaborative tagging (1)
- collection types (1)
- color palettes (1)
- communication network (1)
- complex emotions (1)
- complexity dichotomy (1)
- comprehension (1)
- computational design (1)
- computer science (1)
- computer vision (1)
- conceptualization (1)
- concurrency (1)
- concurrent graph rewriting (1)
- conditional functional dependencies (1)
- conditions (1)
- conflicts and dependencies in (1)
- conformance analysis (1)
- conformance checking (1)
- consistency (1)
- context awareness (1)
- continuous integration (1)
- continuous testing (1)
- contract (1)
- contracts (1)
- control (1)
- control resynthesis (1)
- controlled experiment (1)
- coronavirus 2019 (1)
- corporate takeovers (1)
- cost-effectiveness (1)
- cost-utility analysis (1)
- creativity and innovation management (1)
- critical pairs (1)
- crosscutting wrappers (1)
- cryptocurrency exchanges (1)
- cscw (1)
- cyber (1)
- cyber humanistic (1)
- cyber threat intelligence (1)
- cyber-physical systems (1)
- damage detection (1)
- data center management (1)
- data analysis (1)
- data center management (1)
- data cleaning (1)
- data cleansing (1)
- data correctness checking (1)
- data extraction (1)
- data flow correctness (1)
- data migration (1)
- data objects (1)
- data preparation (1)
- data states (1)
- data transformation (1)
- data-driven demand (1)
- database systems (1)
- database technology (1)
- datasets (1)
- deadline propagation (1)
- deduplication (1)
- deep learning (1)
- delay propagation (1)
- denial-of-service attack (1)
- dependable computing (1)
- dependencies (1)
- dependency (1)
- design (1)
- deterministic properties (1)
- deterministic random walk (1)
- deurema modeling language (1)
- development tools (1)
- diabetes (1)
- difference of Gaussians (1)
- differential (1)
- differential privacy (1)
- diffusion (1)
- digital health app (1)
- digital interventions (1)
- digital startup (1)
- digital therapy (1)
- digital whiteboard (1)
- direct manipulation (1)
- direkte Manipulation (1)
- distributed computing (1)
- distributed data-parallel processing (1)
- distributed ledger technology (1)
- distribution algorithm (1)
- dynamic typing (1)
- dynamic consolidation (1)
- dynamic pricing (1)
- dynamic programming languages (1)
- dynamic reconfiguration (1)
- dynamische Programmiersprachen (1)
- dynamische Sprachen (1)
- dynamische Umsortierung (1)
- e-Learning (1)
- e-learning (1)
- e-lecture (1)
- eHealth (1)
- ePA (1)
- ecological momentary assessment (1)
- ecosystems (1)
- eindeutig (1)
- eingebettete Systeme (1)
- electronic health records (1)
- electronic mail (1)
- embedded systems (1)
- embedding (1)
- emotional (1)
- empathy (1)
- empirical studies (1)
- empirische Studien (1)
- end-stage kidney disease (1)
- endogenous (1)
- energy efficiency (1)
- engine (1)
- engineering (1)
- enrichment calculation (1)
- entity alignment (1)
- enumeration complexity (1)
- epistasis (1)
- erfahrbare Medien (1)
- estimation-of-distribution (1)
- estimation-of-distribution algorithm (1)
- ethics (1)
- event abstraction (1)
- event log (1)
- events (1)
- evolution (1)
- evolution in MDE (1)
- exact exponential-time algorithms (1)
- executable semantics (1)
- experience report (1)
- experimental design (1)
- experiments (1)
- expertise (1)
- explainability (1)
- explainability-accuracy trade-off (1)
- explainable AI (1)
- exploratory programming (1)
- expression (1)
- external knowledge bases (1)
- factual (1)
- failure model (1)
- failure profile (1)
- failure profile model (1)
- fair division (1)
- feature selection (1)
- feedback loop modeling (1)
- fehlende Daten (1)
- flow-based bilateral filter (1)
- flux (1)
- focus plus context visualization (1)
- folksonomy (1)
- force-feedback (1)
- forensics (1)
- formal framework (1)
- formal verification (1)
- formal verification methods (1)
- formale Verifikation (1)
- formales Framework (1)
- formalism (1)
- fully concurrent bisimulation (1)
- functional dependencies (1)
- functional languages (1)
- functional lenses (1)
- functional programming (1)
- funktionale Abhängigkeit (1)
- funktionale Programmierung (1)
- future SOC lab (1)
- game theory (1)
- gaming (1)
- ganzheitlich (1)
- gene (1)
- gene selection (1)
- generalization (1)
- generative multi-discriminative networks (1)
- genetic algorithms (1)
- geocoding (1)
- geographic information systems (1)
- geospatial artificial intelligence (1)
- geospatial data (1)
- geospatial digital twins (1)
- gesture (1)
- grammar-based compression (1)
- graph clustering (1)
- graph databases (1)
- graph languages (1)
- graph pattern matching (1)
- graph queries (1)
- graph replacement categories (1)
- graph transformation systems (1)
- graph transformations (1)
- health apps (1)
- health information privacy concern (1)
- health personnel (1)
- healthcare (1)
- heuristic algorithms (1)
- history (1)
- holistic (1)
- home-based studies (1)
- homomorphic encryption (1)
- hospital (1)
- human computer interaction (1)
- human immunodeficiency virus (1)
- hybrid graph-transformation-systems (1)
- hybride Graph-Transformations-Systeme (1)
- hyperbolic geometry (1)
- identity broker (1)
- identity management (1)
- image captioning (1)
- image processing (1)
- imbalanced learning (1)
- immutable values (1)
- implantable medical device (1)
- implants (1)
- in-memory database (1)
- inattentional blindness (1)
- inclusion (1)
- inclusion dependency (1)
- index (1)
- individuals (1)
- inductive invariant checking (1)
- induktives Invariant Checking (1)
- informatics (1)
- information storage and (1)
- inkrementelle Graphmustersuche (1)
- inkrementelles Graph Pattern Matching (1)
- innovation (1)
- innovation capabilities (1)
- innovation management (1)
- input accuracy (1)
- interaction (1)
- interactive editing (1)
- interactive simulation (1)
- interconnect (1)
- interface (1)
- interpretable machine learning (1)
- invariant checking (1)
- invasive aspects (1)
- job (1)
- k-Induktion (1)
- k-induction (1)
- k-inductive invariant checking (1)
- k-inductive invariants (1)
- k-induktive Invarianten (1)
- k-induktives Invariant-Checking (1)
- key discovery (1)
- knowledge building (1)
- knowledge discovery (1)
- knowledge management (1)
- kontinuierliche Integration (1)
- kontinuierliches Testen (1)
- kontrolliertes Experiment (1)
- künstliche Intelligenz (1)
- label-free (1)
- landmarks (1)
- language specification (1)
- languages (1)
- laser cutting (1)
- layered architecture (1)
- lazy transformation (1)
- leadership (1)
- lean startup approach (1)
- learning (1)
- link discovery (1)
- linked data (1)
- live programming (1)
- local confluence (1)
- location-based (1)
- logic (1)
- logic synthesis (1)
- low back pain (1)
- low-code development approaches (1)
- lower bound (1)
- mHealth (1)
- main memory computing (1)
- management (1)
- many-core (1)
- map reduce (1)
- map/reduce (1)
- maschinelles Lernen (1)
- mass isotopologue distribution (1)
- matching dependencies (1)
- maximal structuring (1)
- mean-variance optimization (1)
- medical (1)
- medical identity theft (1)
- medical malpractice (1)
- mehrdimensionale Belangtrennung (1)
- memory optimization (1)
- memory-based clustering (1)
- memory-based correlation (1)
- memory-based databases (1)
- mental disorders (1)
- mental health (1)
- metadata discovery (1)
- metadata quality (1)
- methodologie (1)
- metric learning (1)
- mixed-methods (1)
- mobile (1)
- mobile application (1)
- mobile devices (1)
- mobile health (1)
- mobile phone (1)
- model generation (1)
- model interpreter (1)
- model verification (1)
- model-based prototyping (1)
- model-driven (1)
- model-driven software engineering (1)
- modelgetriebene Entwicklung (1)
- modeling language (1)
- modellgetriebene Softwareentwicklung (1)
- modelling (1)
- models at runtime (1)
- modular counting (1)
- modularity (1)
- molecular tumor board (1)
- monitoring (1)
- morphic (1)
- mortality (1)
- motion analysis (1)
- motion capture (1)
- motivations (1)
- multi-dimensional separation of concerns (1)
- multi-instances (1)
- multi-perspective visualization (1)
- multi-product pricing (1)
- multilevel systems (1)
- multimodal representations (1)
- multimodal sensing (1)
- multiview classification (1)
- mutation (1)
- mutli-task learning (1)
- network creation games (1)
- networks (1)
- neural (1)
- non-repudiation (1)
- object life cycle synchronization (1)
- object-constraint programming (1)
- one-time password (1)
- openHPI (1)
- organizational change (1)
- organizations (1)
- orthopedic (1)
- orthopedic; (1)
- orts-basiert (1)
- ownership (1)
- pain management (1)
- panorama (1)
- parallel (1)
- parallel computing (1)
- paralleles Rechnen (1)
- parameterized complexity (1)
- parkinson's disease (1)
- parsimonious reduction (1)
- partial application conditions (1)
- partielle Anwendungsbedingungen (1)
- periodic tasks (1)
- periodische Aufgaben (1)
- personal electronic health records (1)
- personal satisfaction (1)
- personalized medicine (1)
- petri net (1)
- phishing (1)
- physical activity assessment (1)
- point clouds (1)
- portability (1)
- power-law (1)
- prefetching (1)
- presentation (1)
- prevention (1)
- price of anarchy (1)
- pricing (1)
- prior knowledge (1)
- probabilistic (1)
- probabilistic models (1)
- probabilistic timed automata (1)
- probabilistische zeitbehaftete Automaten (1)
- process (1)
- process and data integration (1)
- process automation (1)
- process elicitation (1)
- process instance (1)
- process model search (1)
- process modeling (1)
- process refinement (1)
- process scheduling (1)
- process-aware digital twin cockpit (1)
- process-awareness (1)
- processes (1)
- processing (1)
- processor hardware (1)
- profiling (1)
- program (1)
- program analysis (1)
- programming language (1)
- programs (1)
- proportional division (1)
- proteomics (1)
- protocols (1)
- prototypes (1)
- public health medicine (1)
- quantitative analysis (1)
- query matching (1)
- query optimisation (1)
- query rewriting (1)
- querying (1)
- random I (1)
- random forest (1)
- random graphs (1)
- ranking (1)
- rapid prototyping (1)
- reactive (1)
- reaktive Programmierung (1)
- real-time (1)
- real-time systems (1)
- recursive tuning (1)
- reflection (1)
- reinforcement learning (1)
- relational model transformation (1)
- relationale Modelltransformationen (1)
- reliability (1)
- remodularization (1)
- remote collaboration (1)
- remote monitoring (1)
- requirements (1)
- research challenges (1)
- resilient architectures (1)
- resource management (1)
- resource optimization (1)
- restoration (1)
- restricted edges (1)
- retrieval (1)
- reusable aspects (1)
- reuse (1)
- reward (1)
- robustness (1)
- rotor-router model (1)
- runtime adaptations (1)
- runtime analysis (1)
- runtime behavior (1)
- runtime models (1)
- s/t-pattern sequences (1)
- satisfiabilitiy solving (1)
- scale-free networks (1)
- school (environment) (1)
- science mapping (1)
- search plan generation (1)
- security chaos engineering (1)
- security policies (1)
- security risk assessment (1)
- segmentation (1)
- self-adaptive software (1)
- self-driving (1)
- self-learning scheduler (1)
- semantic (1)
- semantic analysis (1)
- semantics preservation (1)
- sensor (1)
- sensor data (1)
- separation of concerns (1)
- service-oriented (1)
- severe acute respiratory (1)
- signal transition graph (1)
- similarity (1)
- similarity learning (1)
- simulator (1)
- single vertex discrepancy (1)
- single-case experimental design (1)
- small files (1)
- smallest grammar problem (1)
- smalltalk (1)
- smart card (1)
- sociology (1)
- software (1)
- software analysis (1)
- software architecture (1)
- software development (1)
- software development processes (1)
- software evolution (1)
- software maintenance (1)
- software product lines (1)
- software tests (1)
- software visualization (1)
- software/instrumentation (1)
- spamming (1)
- speed independence (1)
- speed independent (1)
- stable matchings (1)
- staging (1)
- standards (1)
- static analysis (1)
- statische Analyse (1)
- statistics (1)
- statutes and laws (1)
- steganography (1)
- straight-line (1)
- structured process model (1)
- study (1)
- style description languages (1)
- stylization (1)
- substitution effects (1)
- supervised machine learning (1)
- surveillance (1)
- symbolic graph transformation (1)
- synchronization (1)
- syndrome coronavirus 2 (1)
- synonym discovery (1)
- system of systems (1)
- systems (1)
- t.BPM (1)
- tableau method (1)
- tangible media (1)
- teamwork (1)
- technology adoption (1)
- technology pivot (1)
- tele-TASK (1)
- terrain models (1)
- test-driven fault navigation (1)
- threshold cryptography (1)
- tools (1)
- topics (1)
- tort law (1)
- touch input (1)
- traceability (1)
- tracing (1)
- transfer learning (1)
- transformation (1)
- transformation level (1)
- transformation sequences (1)
- transversal hypergraph (1)
- tree conjecture (1)
- triple graph grammars (1)
- trust (1)
- trust model (1)
- tuple spaces (1)
- typed graph transformation systems (1)
- typography (1)
- unequal shares (1)
- unique (1)
- use-cases (1)
- user interaction (1)
- utility (1)
- verschachtelte Anwednungsbedingungen (1)
- verschachtelte Graphbedingungen (1)
- versioning (1)
- verteilte Datenbanken (1)
- video analysis (1)
- video metadata (1)
- view maintenance (1)
- views (1)
- virtual 3D city models (1)
- virtual groups (1)
- virtual reality (1)
- virtualisierte IT-Infrastruktur (1)
- virtuelle 3D-Stadtmodelle (1)
- visually impaired (1)
- vulnerabilities (1)
- wearable movement sensor (1)
- web application (1)
- web-applications (1)
- weight (1)
- word clouds (1)
- word sense disambiguation (1)
- workflow (1)
- workload prediction (1)
- zero-power defense (1)
- zuverlässige Datenverarbeitung (1)
- zuverlässigen Datenverarbeitung (1)
- Ähnlichkeit (1)
- Ähnlichkeitsmaße (1)
- Ähnlichkeitssuche (1)
- Änderbarkeit (1)
- Übereinstimmungsanalyse (1)
- Überwachung (1)
Institute
- Hasso-Plattner-Institut für Digital Engineering gGmbH (361) (remove)
Linked Open Data (LOD) comprises very many and often large public data sets and knowledge bases. Those datasets are mostly presented in the RDF triple structure of subject, predicate, and object, where each triple represents a statement or fact. Unfortunately, the heterogeneity of available open data requires significant integration steps before it can be used in applications. Meta information, such as ontological definitions and exact range definitions of predicates, are desirable and ideally provided by an ontology. However in the context of LOD, ontologies are often incomplete or simply not available. Thus, it is useful to automatically generate meta information, such as ontological dependencies, range definitions, and topical classifications. Association rule mining, which was originally applied for sales analysis on transactional databases, is a promising and novel technique to explore such data. We designed an adaptation of this technique for min-ing Rdf data and introduce the concept of “mining configurations”, which allows us to mine RDF data sets in various ways. Different configurations enable us to identify schema and value dependencies that in combination result in interesting use cases. To this end, we present rule-based approaches for auto-completion, data enrichment, ontology improvement, and query relaxation. Auto-completion remedies the problem of inconsistent ontology usage, providing an editing user with a sorted list of commonly used predicates. A combination of different configurations step extends this approach to create completely new facts for a knowledge base. We present two approaches for fact generation, a user-based approach where a user selects the entity to be amended with new facts and a data-driven approach where an algorithm discovers entities that have to be amended with missing facts. As knowledge bases constantly grow and evolve, another approach to improve the usage of RDF data is to improve existing ontologies. Here, we present an association rule based approach to reconcile ontology and data. Interlacing different mining configurations, we infer an algorithm to discover synonymously used predicates. Those predicates can be used to expand query results and to support users during query formulation. We provide a wide range of experiments on real world datasets for each use case. The experiments and evaluations show the added value of association rule mining for the integration and usability of RDF data and confirm the appropriateness of our mining configuration methodology.
Unique column combinations of a relational database table are sets of columns that contain only unique values. Discovering such combinations is a fundamental research problem and has many different data management and knowledge discovery applications. Existing discovery algorithms are either brute force or have a high memory load and can thus be applied only to small datasets or samples. In this paper, the wellknown GORDIAN algorithm and "Apriori-based" algorithms are compared and analyzed for further optimization. We greatly improve the Apriori algorithms through efficient candidate generation and statistics-based pruning methods. A hybrid solution HCAGORDIAN combines the advantages of GORDIAN and our new algorithm HCA, and it significantly outperforms all previous work in many situations.
In clinical settings, significant resources are spent on data collection and monitoring patients' health parameters to improve decision-making and provide better care. With increased digitization, the healthcare sector is shifting towards implementing digital technologies for data management and in administration. New technologies offer better treatment opportunities and streamline clinical workflow, but the complexity can cause ineffectiveness, frustration, and errors. To address this, we believe digital solutions alone are not sufficient. Therefore, we take a human-centred design approach for AI development, and apply systems engineering methods to identify system leverage points. We demonstrate how automation enables monitoring clinical parameters, using existing non-intrusive sensor technology, resulting in more resources toward patient care. Furthermore, we provide a framework on digitization of clinical data for integration with data management.
Background:
Childhood and adolescence are critical stages of life for mental health and well-being. Schools are a key setting for mental health promotion and illness prevention. One in five children and adolescents have a mental disorder, about half of mental disorders beginning before the age of 14. Beneficial and explainable artificial intelligence can replace current paper- based and online approaches to school mental health surveys. This can enhance data acquisition, interoperability, data driven analysis, trust and compliance. This paper presents a model for using chatbots for non-obtrusive data collection and supervised machine learning models for data analysis; and discusses ethical considerations pertaining to the use of these models.
Methods:
For data acquisition, the proposed model uses chatbots which interact with students. The conversation log acts as the source of raw data for the machine learning. Pre-processing of the data is automated by filtering for keywords and phrases.
Existing survey results, obtained through current paper-based data collection methods, are evaluated by domain experts (health professionals). These can be used to create a test dataset to validate the machine learning models. Supervised learning
can then be deployed to classify specific behaviour and mental health patterns.
Results:
We present a model that can be used to improve upon current paper-based data collection and manual data analysis methods. An open-source GitHub repository contains necessary tools and components of this model. Privacy is respected through
rigorous observance of confidentiality and data protection requirements. Critical reflection on these ethics and law aspects is included in the project.
Conclusions:
This model strengthens mental health surveillance in schools. The same tools and components could be applied to other public health data. Future extensions of this model could also incorporate unsupervised learning to find clusters and patterns
of unknown effects.
Extract-Transform-Load (ETL) tools are used for the creation, maintenance, and evolution of data warehouses, data marts, and operational data stores. ETL workflows populate those systems with data from various data sources by specifying and executing a DAG of transformations. Over time, hundreds of individual workflows evolve as new sources and new requirements are integrated into the system. The maintenance and evolution of large-scale ETL systems requires much time and manual effort. A key problem is to understand the meaning of unfamiliar attribute labels in source and target databases and ETL transformations. Hard-to-understand attribute labels lead to frustration and time spent to develop and understand ETL workflows. We present a schema decryption technique to support ETL developers in understanding cryptic schemata of sources, targets, and ETL transformations. For a given ETL system, our recommender-like approach leverages the large number of mapped attribute labels in existing ETL workflows to produce good and meaningful decryptions. In this way we are able to decrypt attribute labels consisting of a number of unfamiliar few-letter abbreviations, such as UNP_PEN_INT, which we can decrypt to UNPAID_PENALTY_INTEREST. We evaluate our schema decryption approach on three real-world repositories of ETL workflows and show that our approach is able to suggest high-quality decryptions for cryptic attribute labels in a given schema.
Selection of initial points, the number of clusters and finding proper clusters centers are still the main challenge in clustering processes. In this paper, we suggest genetic algorithm based method which searches several solution spaces simultaneously. The solution spaces are population groups consisting of elements with similar structure. Elements in a group have the same size, while elements in different groups are of different sizes. The proposed algorithm processes the population in groups of chromosomes with one gene, two genes to k genes. These genes hold corresponding information about the cluster centers. In the proposed method, the crossover and mutation operators can accept parents with different sizes; this can lead to versatility in population and information transfer among sub-populations. We implemented the proposed method and evaluated its performance against some random datasets and the Ruspini dataset as well. The experimental results show that the proposed method could effectively determine the appropriate number of clusters and recognize their centers. Overall this research implies that using heterogeneous population in the genetic algorithm can lead to better results.
In order to achieve their business goals, organizations heavily rely on the operational excellence of their business processes. In traditional scenarios, business processes are usually well-structured, clearly specifying when and how certain tasks have to be executed. Flexible and knowledge-intensive processes are gathering momentum, where a knowledge worker drives the execution of a process case and determines the exact process path at runtime. In the case of an exception, the knowledge worker decides on an appropriate handling. While there is initial work on exception handling in well-structured business processes, exceptions in case management have not been sufficiently researched. This paper proposes an exception handling framework for stage-oriented case management languages, namely Guard Stage Milestone Model, Case Management Model and Notation, and Fragment-based Case Management. The effectiveness of the framework is evaluated with two real-world use cases showing that it covers all relevant exceptions and proposed handling strategies.
Program behavior that relies on contextual information, such as physical location or network accessibility, is common in today's applications, yet its representation is not sufficiently supported by programming languages. With context-oriented programming (COP), such context-dependent behavioral variations can be explicitly modularized and dynamically activated. In general, COP could be used to manage any context-specific behavior. However, its contemporary realizations limit the control of dynamic adaptation. This, in turn, limits the interaction of COP's adaptation mechanisms with widely used architectures, such as event-based, mobile, and distributed programming. The JCop programming language extends Java with language constructs for context-oriented programming and additionally provides a domain-specific aspect language for declarative control over runtime adaptations. As a result, these redesigned implementations are more concise and better modularized than their counterparts using plain COP. JCop's main features have been described in our previous publications. However, a complete language specification has not been presented so far. This report presents the entire JCop language including the syntax and semantics of its new language constructs.
Companies develop process models to explicitly describe their business operations. In the same time, business operations, business processes, must adhere to various types of compliance requirements. Regulations, e.g., Sarbanes Oxley Act of 2002, internal policies, best practices are just a few sources of compliance requirements. In some cases, non-adherence to compliance requirements makes the organization subject to legal punishment. In other cases, non-adherence to compliance leads to loss of competitive advantage and thus loss of market share. Unlike the classical domain-independent behavioral correctness of business processes, compliance requirements are domain-specific. Moreover, compliance requirements change over time. New requirements might appear due to change in laws and adoption of new policies. Compliance requirements are offered or enforced by different entities that have different objectives behind these requirements. Finally, compliance requirements might affect different aspects of business processes, e.g., control flow and data flow. As a result, it is infeasible to hard-code compliance checks in tools. Rather, a repeatable process of modeling compliance rules and checking them against business processes automatically is needed. This thesis provides a formal approach to support process design-time compliance checking. Using visual patterns, it is possible to model compliance requirements concerning control flow, data flow and conditional flow rules. Each pattern is mapped into a temporal logic formula. The thesis addresses the problem of consistency checking among various compliance requirements, as they might stem from divergent sources. Also, the thesis contributes to automatically check compliance requirements against process models using model checking. We show that extra domain knowledge, other than expressed in compliance rules, is needed to reach correct decisions. In case of violations, we are able to provide a useful feedback to the user. The feedback is in the form of parts of the process model whose execution causes the violation. In some cases, our approach is capable of providing automated remedy of the violation.
A business process is a set of steps designed to be executed in a certain order to achieve a business value. Such processes are often driven by and documented using process models. Nowadays, process models are also applied to drive process execution. Thus, correctness of business process models is a must. Much of the work has been devoted to check general, domain-independent correctness criteria, such as soundness. However, business processes must also adhere to and show compliance with various regulations and constraints, the so-called compliance requirements. These are domain-dependent requirements.
In many situations, verifying compliance on a model level is of great value, since violations can be resolved in an early stage prior to execution. However, this calls for using formal verification techniques, e.g., model checking, that are too complex for business experts to apply. In this paper, we utilize a visual language. BPMN-Q to express compliance requirements visually in a way similar to that used by business experts to build process models. Still, using a pattern based approach, each BPMN-Qgraph has a formal temporal logic expression in computational tree logic (CTL). Moreover, the user is able to express constraints, i.e., compliance rules, regarding control flow and data flow aspects. In order to provide valuable feedback to a user in case of violations, we depend on temporal logic querying approaches as well as BPMN-Q to visually highlight paths in a process model whose execution causes violations.
Network Topology Discovery and Inventory Listing are two of the primary features of modern network monitoring systems (NMS). Current NMSs rely heavily on active scanning techniques for discovering and mapping network information. Although this approach works, it introduces some major drawbacks such as the performance impact it can exact, specially in larger network environments. As a consequence, scans are often run less frequently which can result in stale information being presented and used by the network monitoring system. Alternatively, some NMSs rely on their agents being deployed on the hosts they monitor. In this article, we present a new approach to Network Topology Discovery and Network Inventory Listing using only passive monitoring and scanning techniques. The proposed techniques rely solely on the event logs produced by the hosts and network devices present within a network. Finally, we discuss some of the advantages and disadvantages of our approach.
Nowadays, business processes are increasingly supported by IT services that produce massive amounts of event data during process execution. Aiming at a better process understanding and improvement, this event data can be used to analyze processes using process mining techniques. Process models can be automatically discovered and the execution can be checked for conformance to specified behavior. Moreover, existing process models can be enhanced and annotated with valuable information, for example for performance analysis. While the maturity of process mining algorithms is increasing and more tools are entering the market, process mining projects still face the problem of different levels of abstraction when comparing events with modeled business activities. Mapping the recorded events to activities of a given process model is essential for conformance checking, annotation and understanding of process discovery results. Current approaches try to abstract from events in an automated way that does not capture the required domain knowledge to fit business activities. Such techniques can be a good way to quickly reduce complexity in process discovery. Yet, they fail to enable techniques like conformance checking or model annotation, and potentially create misleading process discovery results by not using the known business terminology.
In this thesis, we develop approaches that abstract an event log to the same level that is needed by the business. Typically, this abstraction level is defined by a given process model. Thus, the goal of this thesis is to match events from an event log to activities in a given process model. To accomplish this goal, behavioral and linguistic aspects of process models and event logs as well as domain knowledge captured in existing process documentation are taken into account to build semiautomatic matching approaches. The approaches establish a pre--processing for every available process mining technique that produces or annotates a process model, thereby reducing the manual effort for process analysts. While each of the presented approaches can be used in isolation, we also introduce a general framework for the integration of different matching approaches.
The approaches have been evaluated in case studies with industry and using a large industry process model collection and simulated event logs. The evaluation demonstrates the effectiveness and efficiency of the approaches and their robustness towards nonconforming execution logs.
The engineering of digital twins and their user interaction parts with explicated processes, namely processaware digital twin cockpits (PADTCs), is challenging due to the complexity of the systems and the need for information from different disciplines within the engineering process. Therefore, it is interesting to investigate how to facilitate their engineering by using already existing data, namely event logs, and reducing the number of manual steps for their engineering. Current research lacks systematic, automated approaches to derive process-aware digital twin cockpits even though some helpful techniques already exist in the areas of process mining and software engineering. Within this paper, we present a low-code development approach that reduces the amount of hand-written code needed and uses process mining techniques to derive PADTCs. We describe what models could be derived from event log data, which generative steps are needed for the engineering of PADTCs, and how process mining could be incorporated into the resulting application. This process is evaluated using the MIMIC III dataset for the creation of a PADTC prototype for an automated hospital transportation system. This approach can be used for early prototyping of PADTCs as it needs no hand-written code in the first place, but it still allows for the iterative evolvement of the application. This empowers domain experts to create their PADTC prototypes.
In recent years, the increased interest in application areas such as social networks has resulted in a rising popularity of graph-based approaches for storing and processing large amounts of interconnected data. To extract useful information from the growing network structures, efficient querying techniques are required.
In this paper, we propose an approach for graph pattern matching that allows a uniform handling of arbitrary constraints over the query vertices. Our technique builds on a previously introduced matching algorithm, which takes concrete host graph information into account to dynamically adapt the employed search plan during query execution. The dynamic algorithm is combined with an existing static approach for search plan generation, resulting in a hybrid technique which we further extend by a more sophisticated handling of filtering effects caused by constraint checks. We evaluate the presented concepts empirically based on an implementation for our graph pattern matching tool, the Story Diagram Interpreter, with queries and data provided by the LDBC Social Network Benchmark. Our results suggest that the hybrid technique may improve search efficiency in several cases, and rarely reduces efficiency.
LoANs
(2019)
Recently, deep neural networks have achieved remarkable performance on the task of object detection and recognition. The reason for this success is mainly grounded in the availability of large scale, fully annotated datasets, but the creation of such a dataset is a complicated and costly task. In this paper, we propose a novel method for weakly supervised object detection that simplifies the process of gathering data for training an object detector. We train an ensemble of two models that work together in a student-teacher fashion. Our student (localizer) is a model that learns to localize an object, the teacher (assessor) assesses the quality of the localization and provides feedback to the student. The student uses this feedback to learn how to localize objects and is thus entirely supervised by the teacher, as we are using no labels for training the localizer. In our experiments, we show that our model is very robust to noise and reaches competitive performance compared to a state-of-the-art fully supervised approach. We also show the simplicity of creating a new dataset, based on a few videos (e.g. downloaded from YouTube) and artificially generated data.
Data integration aims to combine data of different sources and to provide users with a unified view on these data. This task is as challenging as valuable. In this thesis we propose algorithms for dependency discovery to provide necessary information for data integration. We focus on inclusion dependencies (INDs) in general and a special form named conditional inclusion dependencies (CINDs): (i) INDs enable the discovery of structure in a given schema. (ii) INDs and CINDs support the discovery of cross-references or links between schemas. An IND “A in B” simply states that all values of attribute A are included in the set of values of attribute B. We propose an algorithm that discovers all inclusion dependencies in a relational data source. The challenge of this task is the complexity of testing all attribute pairs and further of comparing all of each attribute pair's values. The complexity of existing approaches depends on the number of attribute pairs, while ours depends only on the number of attributes. Thus, our algorithm enables to profile entirely unknown data sources with large schemas by discovering all INDs. Further, we provide an approach to extract foreign keys from the identified INDs. We extend our IND discovery algorithm to also find three special types of INDs: (i) Composite INDs, such as “AB in CD”, (ii) approximate INDs that allow a certain amount of values of A to be not included in B, and (iii) prefix and suffix INDs that represent special cross-references between schemas. Conditional inclusion dependencies are inclusion dependencies with a limited scope defined by conditions over several attributes. Only the matching part of the instance must adhere the dependency. We generalize the definition of CINDs distinguishing covering and completeness conditions and define quality measures for conditions. We propose efficient algorithms that identify covering and completeness conditions conforming to given quality thresholds. The challenge for this task is twofold: (i) Which (and how many) attributes should be used for the conditions? (ii) Which attribute values should be chosen for the conditions? Previous approaches rely on pre-selected condition attributes or can only discover conditions applying to quality thresholds of 100%. Our approaches were motivated by two application domains: data integration in the life sciences and link discovery for linked open data. We show the efficiency and the benefits of our approaches for use cases in these domains.
Data dependencies, or integrity constraints, are used to improve the quality of a database schema, to optimize queries, and to ensure consistency in a database. In the last years conditional dependencies have been introduced to analyze and improve data quality. In short, a conditional dependency is a dependency with a limited scope defined by conditions over one or more attributes. Only the matching part of the instance must adhere to the dependency. In this paper we focus on conditional inclusion dependencies (CINDs). We generalize the definition of CINDs, distinguishing covering and completeness conditions. We present a new use case for such CINDs showing their value for solving complex data quality tasks. Further, we define quality measures for conditions inspired by precision and recall. We propose efficient algorithms that identify covering and completeness conditions conforming to given quality thresholds. Our algorithms choose not only the condition values but also the condition attributes automatically. Finally, we show that our approach efficiently provides meaningful and helpful results for our use case.