004 Datenverarbeitung; Informatik
Refine
Year of publication
Document Type
- Article (40)
- Monograph/Edited Volume (33)
- Doctoral Thesis (32)
- Postprint (5)
- Conference Proceeding (2)
- Part of a Book (1)
- Other (1)
Language
- English (114) (remove)
Keywords
- machine learning (8)
- maschinelles Lernen (5)
- Smalltalk (4)
- cyber-physical systems (4)
- probabilistic timed systems (4)
- qualitative Analyse (4)
- qualitative analysis (4)
- quantitative Analyse (4)
- quantitative analysis (4)
- Cloud Computing (3)
- Forschungsprojekte (3)
- Future SOC Lab (3)
- In-Memory Technologie (3)
- Multicore Architekturen (3)
- artifical intelligence (3)
- blockchain (3)
- business process management (3)
- cloud computing (3)
- multicore architectures (3)
- research projects (3)
- smart contracts (3)
- 3D visualization (2)
- 3D-Visualisierung (2)
- Anomalieerkennung (2)
- Bounded Model Checking (2)
- Datenaufbereitung (2)
- Digitalisierung (2)
- Forschungskolleg (2)
- Graphentransformationssysteme (2)
- Hasso Plattner Institute (2)
- Hasso-Plattner-Institut (2)
- Identitätsmanagement (2)
- In-Memory technology (2)
- Klausurtagung (2)
- MOOC (2)
- Modellprüfung (2)
- Ph.D. retreat (2)
- Service-oriented Systems Engineering (2)
- Sicherheit (2)
- Versionsverwaltung (2)
- Werkzeuge (2)
- anomaly detection (2)
- bounded model checking (2)
- business processes (2)
- classification (2)
- cloud (2)
- clustering (2)
- computer vision (2)
- cyber-physische Systeme (2)
- data preparation (2)
- deep learning (2)
- deferred choice (2)
- digital education (2)
- digitalization (2)
- formal semantics (2)
- functional dependencies (2)
- funktionale Abhängigkeiten (2)
- geospatial data (2)
- graph transformation systems (2)
- identity management (2)
- inclusion dependencies (2)
- intrusion detection (2)
- künstliche Intelligenz (2)
- maschinelles Sehen (2)
- memory (2)
- mobile mapping (2)
- model checking (2)
- nested graph conditions (2)
- openHPI (2)
- oracles (2)
- outlier detection (2)
- probabilistische gezeitete Systeme (2)
- probabilistische zeitgesteuerte Systeme (2)
- research school (2)
- security (2)
- self-sovereign identity (2)
- service-oriented systems engineering (2)
- tiefes Lernen (2)
- typed attributed graphs (2)
- user-generated content (2)
- version control (2)
- workflow patterns (2)
- 0-day (1)
- 3D city model (1)
- 3D geovisualization (1)
- 3D point cloud (1)
- 3D point clouds (1)
- 3D portrayal (1)
- 3D-Geovisualisierung (1)
- 3D-Punktwolke (1)
- 3D-Punktwolken (1)
- 3D-Rendering (1)
- 3D-Stadtmodell (1)
- ACINQ (1)
- APT (1)
- ASIC (1)
- Activity-oriented Optimization (1)
- Advanced Persistent Threats (1)
- Agile (1)
- Agilität (1)
- Aktivitäten (1)
- Algebraic methods (1)
- Ambiguity (1)
- Ambiguität (1)
- Analog-zu-Digital-Konvertierung (1)
- Anfrageoptimierung (1)
- Angriffserkennung (1)
- Architekturadaptation (1)
- Archivanalyse (1)
- Arzt-Patient-Beziehung (1)
- Attributsicherung (1)
- Ausreißererkennung (1)
- Australian securities exchange (1)
- Auswirkungen (1)
- BCCC (1)
- BPMN (1)
- BTC (1)
- Bahnwesen (1)
- Bedrohungserkennung (1)
- Behavior change (1)
- Betriebssysteme (1)
- Big Data (1)
- Big Five model (1)
- Bildverarbeitung (1)
- BitShares (1)
- Bitcoin Core (1)
- Blockchain (1)
- Blockchain Auth (1)
- Blockchain-Konsortium R3 (1)
- Blockkette (1)
- Blockstack (1)
- Blockstack ID (1)
- Blumix-Plattform (1)
- Blöcke (1)
- Bounded Backward Model Checking (1)
- Business Process Management (1)
- Business process modeling (1)
- Byzantine Agreement (1)
- CityGML (1)
- Clinical predictive modeling (1)
- Clusteranalyse (1)
- Clustering (1)
- Colored Coins (1)
- Compound Values (1)
- Computational Photography (1)
- Computing (1)
- Conceptual modeling (1)
- Creative (1)
- Cyber-Sicherheit (1)
- Cyber-physikalische Systeme (1)
- DAO (1)
- DBMS (1)
- DPoS (1)
- Data Profiling (1)
- Data Structure Optimization (1)
- Data modeling (1)
- Data profiling (1)
- Data warehouse (1)
- Data-Mining (1)
- Data-Science (1)
- Dateistruktur (1)
- Datenanalyse (1)
- Datenbank (1)
- Datenbanksysteme (1)
- Datenintegration (1)
- Datenmodelle (1)
- Datenqualität (1)
- Datensatz (1)
- Datenstrukturoptimierung (1)
- Datensynthese (1)
- Datentransformation (1)
- Datenvisualisierung (1)
- Debugging (1)
- Decision support (1)
- Dekubitus (1)
- Delegated Proof-of-Stake (1)
- Denkweise (1)
- Design Thinking (1)
- Design-Forschung (1)
- Direkte Manipulation (1)
- Distributed Proof-of-Research (1)
- Distributed-Ledger-Technologie (DLT) (1)
- Domänenspezifische Modellierung (1)
- Dubletten (1)
- Duplikaterkennung (1)
- E-Learning (1)
- E-Wallet (1)
- ECDSA (1)
- Echtzeit (1)
- Echtzeit-Rendering (1)
- Einbruchserkennung (1)
- Endpunktsicherheit (1)
- Entitätsverknüpfung (1)
- Entscheidungsfindung (1)
- Entscheidungsmanagement (1)
- Entscheidungsmodelle (1)
- Enumeration algorithm (1)
- Eris (1)
- Erkennung von Metadaten (1)
- Ether (1)
- Ethereum (1)
- European Union (1)
- Europäische Union (1)
- Evaluation (1)
- Exploration (1)
- Feature selection (1)
- Federated Byzantine Agreement (1)
- Fehlertoleranz (1)
- Fernerkundung (1)
- FollowMyVote (1)
- Fork (1)
- Formal modelling (1)
- GPU (1)
- GPU acceleration (1)
- GPU-Beschleunigung (1)
- Gaussian process state-space models (1)
- Gaussian processes (1)
- Gauß-Prozess Zustandsraummodelle (1)
- Gauß-Prozesse (1)
- Gene expression (1)
- General Earth and Planetary Sciences (1)
- Generalized Discrimination Networks (1)
- Geodaten (1)
- Geography, Planning and Development (1)
- Geschäftsprozessarchitekturen (1)
- Geschäftsprozessmanagement (1)
- Gewinnung benannter Entitäten (1)
- GitHub (1)
- GraalVM (1)
- Graph logic (1)
- Graph-Mining (1)
- Graphableitung (1)
- Graphreparatur (1)
- Gridcoin (1)
- HPI Schul-Cloud (1)
- Hard Fork (1)
- Hashed Timelock Contracts (1)
- Hasserkennung (1)
- Heuristiken (1)
- Hyrise (1)
- Häkeln (1)
- IDS (1)
- Ideation (1)
- Ideenfindung (1)
- Identität (1)
- Impact (1)
- Implementation in Organizations (1)
- Implementierung in Organisationen (1)
- In-Memory (1)
- Informationsextraktion (1)
- Inklusionsabhängigkeiten (1)
- Innovation (1)
- Interdisciplinary Teams (1)
- Internet der Dinge (1)
- Internet of Things (1)
- Interpretability (1)
- Interpreter (1)
- Interval Timed Automata (1)
- IoT (1)
- Japanese Blockchain Consortium (1)
- Japanisches Blockchain-Konsortium (1)
- Java (1)
- Karten (1)
- Kausalität (1)
- Kette (1)
- Klassifikation (1)
- Klassifizierung (1)
- Konsensalgorithmus (1)
- Konsensprotokoll (1)
- Konsensprotokolle (1)
- Konsistenzrestauration (1)
- Kreativität (1)
- Kunstanalyse (1)
- Künstliche Intelligenz (1)
- Laserscanning (1)
- Laufzeitmodelle (1)
- Learning Analytics (1)
- Lebendigkeit (1)
- Leistungsmodelle von virtuellen Maschinen (1)
- LiDAR (1)
- Lightning Network (1)
- Live-Migration (1)
- Live-Programmierung (1)
- Lively Kernel (1)
- Lock-Time-Parameter (1)
- Lösungsraum (1)
- MERLOT (1)
- MOOCs (1)
- Machine-Learning (1)
- Machinelles Lernen (1)
- Maschinelles Lernen (1)
- Maschinen (1)
- Measurement (1)
- Messung (1)
- Metacrate (1)
- Metadaten (1)
- Micropayment-Kanäle (1)
- Microsoft Azur (1)
- Mindset (1)
- Minimal hitting set (1)
- Mobile Mapping (1)
- Mobile-Mapping (1)
- Modelle mit mehreren Versionen (1)
- Modellreparatur (1)
- Multidisciplinary Teams (1)
- NASDAQ (1)
- NameID (1)
- Namecoin (1)
- Nephrology (1)
- Netzwerkprotokolle (1)
- Non-photorealistic Rendering (1)
- Nutzerinteraktion (1)
- Objects (1)
- Objekte (1)
- Off-Chain-Transaktionen (1)
- Onename (1)
- Online Learning Environments (1)
- OpenBazaar (1)
- OptoGait (1)
- Oracles (1)
- Orphan Block (1)
- P2P (1)
- Patientenermündigung (1)
- Peer-to-Peer Netz (1)
- Peercoin (1)
- PoB (1)
- PoS (1)
- PoW (1)
- Posenabschätzung (1)
- Prior knowledge (1)
- Privatsphäre (1)
- Problem Solving (1)
- Problemlösung (1)
- Process (1)
- Process Execution (1)
- Programmieren (1)
- Programmiererlebnis (1)
- Programmierwerkzeuge (1)
- Proof-of-Burn (1)
- Proof-of-Stake (1)
- Proof-of-Work (1)
- Prototyping (1)
- Prozess (1)
- Prozessmodelle (1)
- Psychotherapie (1)
- Python (1)
- Quanten-Computing (1)
- Query-Optimierung (1)
- RL (1)
- Regressionstests (1)
- Reproducible benchmarking (1)
- Resource Allocation (1)
- Resource Management (1)
- Reverse Engineering (1)
- Ripple (1)
- Ruby (1)
- Runtime-monitoring (1)
- SCP (1)
- SHA (1)
- SIEM (1)
- SPV (1)
- SWIRL (1)
- Savanne (1)
- Schriftartgestaltung (1)
- Schriftrendering (1)
- Schwierigkeitsgrad (1)
- Scrollytelling (1)
- Selbst-Adaptive Software (1)
- Self-Regulated Learning (1)
- Sequenzeigenschaften (1)
- Serialisierung (1)
- Service-Oriented Architecture (1)
- Sicherheitsanalyse (1)
- Simplified Payment Verification (1)
- Situationsbewusstsein (1)
- Skalierbarkeit der Blockchain (1)
- Skriptsprachen (1)
- Slock.it (1)
- Soft Fork (1)
- Software-Evolution (1)
- Software/Hardware Co-Design (1)
- Solution Space (1)
- Soziale Medien (1)
- Specification (1)
- Spezifikation von gezeiteten Graph Transformationen (1)
- Sprachlernen im Limes (1)
- Squeak (1)
- Squeak/Smalltalk (1)
- Standardisierung (1)
- Steemit (1)
- Stellar Consensus Protocol (1)
- Storj (1)
- Suchtberatung und -therapie (1)
- Telemedizin (1)
- Temporallogik (1)
- Testergebnisse (1)
- Testpriorisierungs (1)
- Texterkennung (1)
- Textklassifikation (1)
- The Bitfury Group (1)
- The DAO (1)
- Timed Automata (1)
- Tools (1)
- Trajektorien (1)
- Transaktion (1)
- Transversal hypergraph (1)
- Tripel-Graph-Grammatiken (1)
- Two-Way-Peg (1)
- Unique column combination (1)
- Unspent Transaction Output (1)
- User Experience (1)
- VUCA-World (1)
- Validation (1)
- Verbundwerte (1)
- Verhaltensänderung (1)
- Verlässlichkeit (1)
- Vertrauen (1)
- Verträge (1)
- Veränderungsanalyse (1)
- Virtual Machines (1)
- Virtuelle Maschinen (1)
- Visualisierungskonzept-Exploration (1)
- Vorhersage (1)
- W[3]-Completeness (1)
- Water Science and Technology (1)
- Watson IoT (1)
- Wearable (1)
- Werkzeugbau (1)
- Wicked Problems (1)
- Wolke (1)
- Wüstenbildung (1)
- Zebris (1)
- Zielvorgabe (1)
- Zookos Dreieck (1)
- Zookos triangle (1)
- addiction care (1)
- advanced persistent threat (1)
- advanced threats (1)
- agil (1)
- altchain (1)
- alternative chain (1)
- analog-to-digital conversion (1)
- apt (1)
- architectural adaptation (1)
- archive analysis (1)
- art analysis (1)
- asset management (1)
- atomic swap (1)
- attribute assurance (1)
- autonomous (1)
- behaviourally correct learning (1)
- benutzergenerierte Inhalte (1)
- bidirectional payment channels (1)
- bildbasiertes Rendering (1)
- bitcoins (1)
- blockchain consortium (1)
- blockchain-übergreifend (1)
- blocks (1)
- blumix platform (1)
- bounded backward model checking (1)
- brand personality (1)
- business process architectures (1)
- categories (1)
- causal discovery (1)
- causal structure learning (1)
- causality (1)
- chain (1)
- change detection (1)
- code generation (1)
- compositional analysis (1)
- computational photography (1)
- computer-aided design (1)
- computer-mediated therapy (1)
- computervermittelte Therapie (1)
- computing (1)
- confirmation period (1)
- confluence (1)
- consensus algorithm (1)
- consensus protocol (1)
- consensus protocols (1)
- consistency restoration (1)
- consistent learning (1)
- contest period (1)
- continuous integration (1)
- contracts (1)
- convolutional neural networks (1)
- creativity (1)
- crochet (1)
- cross-chain (1)
- cultural heritage (1)
- cumulative culture (1)
- cyber-physikalische Systeme (1)
- cybersecurity (1)
- data analytics (1)
- data dependencies (1)
- data integration (1)
- data mining (1)
- data models (1)
- data pipeline (1)
- data profiling (1)
- data quality (1)
- data science (1)
- data set (1)
- data synthesis (1)
- data visualization (1)
- data wrangling (1)
- data-driven (1)
- database (1)
- database optimization (1)
- database systems (1)
- datengetrieben (1)
- debugging (1)
- decentral identities (1)
- decentralized autonomous organization (1)
- decision management (1)
- decision mining (1)
- decision models (1)
- decubitus (1)
- deduplication (1)
- deep Gaussian processes (1)
- demografische Informationen (1)
- demographic information (1)
- dependability (1)
- desertification (1)
- design research (1)
- dezentrale Identitäten (1)
- dezentrale autonome Organisation (1)
- difficulty (1)
- difficulty target (1)
- digital enlightenment (1)
- digital learning platform (1)
- digital picture archive (1)
- digital sovereignty (1)
- digital whiteboard (1)
- digitale Aufklärung (1)
- digitale Bildung (1)
- digitale Lernplattform (1)
- digitale Souveränität (1)
- digitales Bildarchiv (1)
- digitales Whiteboard (1)
- direct manipulation (1)
- discrete-event model (1)
- diskretes Ereignismodell (1)
- distributed computation (1)
- distributed performance monitoring (1)
- distributed systems (1)
- doctor-patient relationship (1)
- domain-specific modeling (1)
- doppelter Hashwert (1)
- double hashing (1)
- drift theory (1)
- duplicate detection (1)
- dynamic systems (1)
- dynamische Systeme (1)
- e-learning (1)
- electrical muscle stimulation (1)
- elektrische Muskelstimulation (1)
- endpoint security (1)
- entity linking (1)
- entity resolution (1)
- erzeugende gegnerische Netzwerke (1)
- evaluation (1)
- evolutionary computation (1)
- experience (1)
- exploration (1)
- exploratives Programmieren (1)
- exploratory programming (1)
- extend (1)
- fault tolerance (1)
- federated voting (1)
- file structure (1)
- font engineering (1)
- font rendering (1)
- fortschrittliche Angriffe (1)
- gait analysis algorithm (1)
- gefaltete neuronale Netze (1)
- generalized discrimination networks (1)
- generative adversarial networks (1)
- geschichtsbewusste Laufzeit-Modelle (1)
- getypte Attributierte Graphen (1)
- global model management (1)
- globales Modellmanagement (1)
- grammars (1)
- graph inference (1)
- graph mining (1)
- graph repair (1)
- graph-transformations (1)
- hashrate (1)
- hate speech detection (1)
- heuristics (1)
- higher education (1)
- history-aware runtime models (1)
- human-centered (1)
- hybrid systems (1)
- identity (1)
- image processing (1)
- image stylization (1)
- image-based rendering (1)
- immediacy (1)
- in-memory (1)
- in-memory technology (1)
- incremental graph query evaluation (1)
- index selection (1)
- inertial measurement unit (1)
- information extraction (1)
- inkrementelle Ausführung von Graphanfragen (1)
- innovation (1)
- integrated development environments (1)
- integrierte Entwicklungsumgebungen (1)
- intelligente Verträge (1)
- inter-chain (1)
- interactive media (1)
- interaktive Medien (1)
- interdisziplinäre Teams (1)
- interpreters (1)
- interval probabilistic timed systems (1)
- interval probabilistische zeitgesteuerte Systeme (1)
- interval timed automata (1)
- intuitive Benutzeroberflächen (1)
- intuitive interfaces (1)
- invention (1)
- invention mechanism (1)
- juridical recording (1)
- k-inductive invariant checking (1)
- k-induktive Invariantenprüfung (1)
- kausale Entdeckung (1)
- kausales Strukturlernen (1)
- kompositionale Analyse (1)
- konsistentes Lernen (1)
- kontinuierliche Integration (1)
- kulturelles Erbe (1)
- language learning in the limit (1)
- laserscanning (1)
- learning (1)
- lebenslanges Lernen (1)
- lebenszentriert (1)
- ledger assets (1)
- left recursion (1)
- level-replacement systems (1)
- life-centered (1)
- lifelong learning (1)
- live migration (1)
- live programming (1)
- liveness (1)
- machine (1)
- machines (1)
- maps (1)
- maschinelle Verarbeitung natürlicher Sprache (1)
- medical documentation (1)
- medizinische Dokumentation (1)
- mehrsprachige Ausführungsumgebungen (1)
- menschenzentriert (1)
- merged mining (1)
- merkle root (1)
- metacrate (1)
- metadata (1)
- metadata detection (1)
- methods (1)
- metric temporal logic (1)
- metric termporal graph logic (1)
- metrisch temporale Graph Logic (1)
- metrische Temporallogik (1)
- microcredential (1)
- micropayment (1)
- micropayment channels (1)
- miner (1)
- mining (1)
- mining hardware (1)
- minting (1)
- model (1)
- model repair (1)
- model-driven engineering (1)
- model-driven software engineering (1)
- modellgetriebene Entwicklung (1)
- modellgetriebene Softwaretechnik (1)
- multi-version models (1)
- multidisziplinäre Teams (1)
- named entity mining (1)
- natural language processing (1)
- network protocols (1)
- non-photorealistic rendering (1)
- nonce (1)
- novelty detection (1)
- nutzergenerierte Inhalte (1)
- object-oriented programming (1)
- objektorientiertes Programmieren (1)
- off-chain transaction (1)
- online course creation (1)
- online course design (1)
- open innovation (1)
- operating systems (1)
- optical character recognition (1)
- order dependencies (1)
- packrat parsing (1)
- parallel and sequential independence (1)
- parallel processing (1)
- parallele Verarbeitung (1)
- parallele und Sequentielle Unabhängigkeit (1)
- parsing expression grammars (1)
- patent (1)
- patient empowerment (1)
- peer-to-peer network (1)
- pegged sidechains (1)
- performance models of virtual machines (1)
- personality prediction (1)
- polyglot execution environments (1)
- polyglot programming (1)
- polyglottes Programmieren (1)
- pose estimation (1)
- prediction (1)
- privacy (1)
- probabilistic machine learning (1)
- probabilistisches maschinelles Lernen (1)
- process mining (1)
- process models (1)
- programming (1)
- programming experience (1)
- programming tools (1)
- programs (1)
- prototyping (1)
- psychotherapy (1)
- public dataset (1)
- qualitative model (1)
- qualitatives Modell (1)
- quantum computing (1)
- query optimization (1)
- quorum slices (1)
- railways (1)
- real-time (1)
- real-time rendering (1)
- rechnerunterstütztes Konstruieren (1)
- reconfigurable systems (1)
- regression testing (1)
- reinforcement learning (1)
- remote sensing (1)
- reverse engineering (1)
- rootstock (1)
- runtime models (1)
- runtime monitoring (1)
- räumliche Geodaten (1)
- savanna (1)
- scalability of blockchain (1)
- scarce tokens (1)
- schwach überwachtes maschinelles Lernen (1)
- scripting languages (1)
- scrollytelling (1)
- security analytics (1)
- selbst-souveräne Identitäten (1)
- selbstbestimmte Identitäten (1)
- self-adaptive software (1)
- self-driving (1)
- semantic classification (1)
- semantische Klassifizierung (1)
- sequence properties (1)
- serialization (1)
- serverseitiges 3D-Rendering (1)
- serverside 3D rendering (1)
- service-oriented architectures (1)
- serviceorientierte Architekturen (1)
- sidechain (1)
- simulation (1)
- situational awareness (1)
- small talk (1)
- smalltalk (1)
- social media (1)
- social media analysis (1)
- software evolution (1)
- software/hardware co-design (1)
- specification of timed graph transformations (1)
- squeak (1)
- standardization (1)
- stark verhaltenskorrekt sperrend (1)
- static source-code analysis (1)
- statische Quellcodeanalyse (1)
- stochastic process (1)
- strongly behaviourally correct locking (1)
- symbolic analysis (1)
- symbolic graphs (1)
- symbolische Analyse (1)
- symbolische Graphen (1)
- synchronization (1)
- tabellarische Dateien (1)
- tabular data (1)
- technology (1)
- telemedicine (1)
- temporal graph queries (1)
- temporal logic (1)
- temporale Graphanfragen (1)
- test case prioritization (1)
- test results (1)
- text classification (1)
- text mining (1)
- threat detection (1)
- tiefe Gauß-Prozesse (1)
- timed automata (1)
- tool building (1)
- tools (1)
- trajectories (1)
- transaction (1)
- triple graph grammars (1)
- trust (1)
- typisierte attributierte Graphen (1)
- unique column combinations (1)
- unsupervised (1)
- user experience (1)
- user interaction (1)
- variational inference (1)
- variationelle Inferenz (1)
- verhaltenskorrektes Lernen (1)
- verifiable credentials (1)
- verschachtelte Anwendungsbedingungen (1)
- verschachtelte Graphbedingungen (1)
- verteilte Berechnung (1)
- verteilte Leistungsüberwachung (1)
- verzwickte Probleme (1)
- virtual (1)
- virtual machines (1)
- virtual reality (1)
- virtuell (1)
- virtuelle Maschinen (1)
- virtuelle Realität (1)
- visual language (1)
- visualization concept exploration (1)
- visuelle Sprache (1)
- weak supervision (1)
- wearables (1)
- web-based development (1)
- web-based development environment (1)
- web-basierte Entwicklungsumgebung (1)
- webbasierte Entwicklung (1)
- zero-day (1)
- überprüfbare Nachweise (1)
Institute
- Hasso-Plattner-Institut für Digital Engineering GmbH (114) (remove)
If taking a flipped learning approach, MOOC content can be used for online pre-class instruction. After which students can put the knowledge they gained from the MOOC into practice either synchronously or asynchronously. This study examined one such, asynchronous, course in teacher education. The course ran with 40 students over 13 weeks from February to May 2020. A case study approach was followed using mixed methods to assess the efficacy of the course. Quantitative data was gathered on achievement of learning outcomes, online engagement, and satisfaction. Qualitative data was gathered via student interviews from which a thematic analysis was undertaken. From a combined analysis of the data, three themes emerged as pertinent to course efficacy: quality and quantity of communication and collaboration; suitability of the MOOC; and significance for career development.
With recent advances in the area of information extraction, automatically extracting structured information from a vast amount of unstructured textual data becomes an important task, which is infeasible for humans to capture all information manually. Named entities (e.g., persons, organizations, and locations), which are crucial components in texts, are usually the subjects of structured information from textual documents. Therefore, the task of named entity mining receives much attention. It consists of three major subtasks, which are named entity recognition, named entity linking, and relation extraction.
These three tasks build up an entire pipeline of a named entity mining system, where each of them has its challenges and can be employed for further applications. As a fundamental task in the natural language processing domain, studies on named entity recognition have a long history, and many existing approaches produce reliable results. The task is aiming to extract mentions of named entities in text and identify their types. Named entity linking recently received much attention with the development of knowledge bases that contain rich information about entities. The goal is to disambiguate mentions of named entities and to link them to the corresponding entries in a knowledge base. Relation extraction, as the final step of named entity mining, is a highly challenging task, which is to extract semantic relations between named entities, e.g., the ownership relation between two companies.
In this thesis, we review the state-of-the-art of named entity mining domain in detail, including valuable features, techniques, evaluation methodologies, and so on. Furthermore, we present two of our approaches that focus on the named entity linking and relation extraction tasks separately.
To solve the named entity linking task, we propose the entity linking technique, BEL, which operates on a textual range of relevant terms and aggregates decisions from an ensemble of simple classifiers. Each of the classifiers operates on a randomly sampled subset of the above range. In extensive experiments on hand-labeled and benchmark datasets, our approach outperformed state-of-the-art entity linking techniques, both in terms of quality and efficiency.
For the task of relation extraction, we focus on extracting a specific group of difficult relation types, business relations between companies. These relations can be used to gain valuable insight into the interactions between companies and perform complex analytics, such as predicting risk or valuating companies. Our semi-supervised strategy can extract business relations between companies based on only a few user-provided seed company pairs. By doing so, we also provide a solution for the problem of determining the direction of asymmetric relations, such as the ownership_of relation. We improve the reliability of the extraction process by using a holistic pattern identification method, which classifies the generated extraction patterns. Our experiments show that we can accurately and reliably extract new entity pairs occurring in the target relation by using as few as five labeled seed pairs.
The “HPI Future SOC Lab” is a cooperation of the Hasso Plattner Institute (HPI) and industry partners. Its mission is to enable and promote exchange and interaction between the research community and the industry partners.
The HPI Future SOC Lab provides researchers with free of charge access to a complete infrastructure of state of the art hard and software. This infrastructure includes components, which might be too expensive for an ordinary research environment, such as servers with up to 64 cores and 2 TB main memory. The offerings address researchers particularly from but not limited to the areas of computer science and business information systems. Main areas of research include cloud computing, parallelization, and In-Memory technologies.
This technical report presents results of research projects executed in 2017. Selected projects have presented their results on April 25th and November 15th 2017 at the Future SOC Lab Day events.
3D point clouds are a universal and discrete digital representation of three-dimensional objects and environments. For geospatial applications, 3D point clouds have become a fundamental type of raw data acquired and generated using various methods and techniques. In particular, 3D point clouds serve as raw data for creating digital twins of the built environment.
This thesis concentrates on the research and development of concepts, methods, and techniques for preprocessing, semantically enriching, analyzing, and visualizing 3D point clouds for applications around transport infrastructure. It introduces a collection of preprocessing techniques that aim to harmonize raw 3D point cloud data, such as point density reduction and scan profile detection. Metrics such as, e.g., local density, verticality, and planarity are calculated for later use. One of the key contributions tackles the problem of analyzing and deriving semantic information in 3D point clouds. Three different approaches are investigated: a geometric analysis, a machine learning approach operating on synthetically generated 2D images, and a machine learning approach operating on 3D point clouds without intermediate representation.
In the first application case, 2D image classification is applied and evaluated for mobile mapping data focusing on road networks to derive road marking vector data. The second application case investigates how 3D point clouds can be merged with ground-penetrating radar data for a combined visualization and to automatically identify atypical areas in the data. For example, the approach detects pavement regions with developing potholes. The third application case explores the combination of a 3D environment based on 3D point clouds with panoramic imagery to improve visual representation and the detection of 3D objects such as traffic signs.
The presented methods were implemented and tested based on software frameworks for 3D point clouds and 3D visualization. In particular, modules for metric computation, classification procedures, and visualization techniques were integrated into a modular pipeline-based C++ research framework for geospatial data processing, extended by Python machine learning scripts. All visualization and analysis techniques scale to large real-world datasets such as road networks of entire cities or railroad networks.
The thesis shows that some use cases allow taking advantage of established image vision methods to analyze images rendered from mobile mapping data efficiently. The two presented semantic classification methods working directly on 3D point clouds are use case independent and show similar overall accuracy when compared to each other. While the geometry-based method requires less computation time, the machine learning-based method supports arbitrary semantic classes but requires training the network with ground truth data. Both methods can be used in combination to gradually build this ground truth with manual corrections via a respective annotation tool.
This thesis contributes results for IT system engineering of applications, systems, and services that require spatial digital twins of transport infrastructure such as road networks and railroad networks based on 3D point clouds as raw data. It demonstrates the feasibility of fully automated data flows that map captured 3D point clouds to semantically classified models. This provides a key component for seamlessly integrated spatial digital twins in IT solutions that require up-to-date, object-based, and semantically enriched information about the built environment.
CovRadar
(2022)
The ongoing pandemic caused by SARS-CoV-2 emphasizes the importance of genomic surveillance to understand the evolution of the virus, to monitor the viral population, and plan epidemiological responses. Detailed analysis, easy visualization and intuitive filtering of the latest viral sequences are powerful for this purpose. We present CovRadar, a tool for genomic surveillance of the SARS-CoV-2 Spike protein. CovRadar consists of an analytical pipeline and a web application that enable the analysis and visualization of hundreds of thousand sequences. First, CovRadar extracts the regions of interest using local alignment, then builds a multiple sequence alignment, infers variants and consensus and finally presents the results in an interactive app, making accessing and reporting simple, flexible and fast.
Decubitus is one of the most relevant diseases in nursing and the most expensive to treat. It is caused by sustained pressure on tissue, so it particularly affects bed-bound patients. This work lays a foundation for pressure mattress-based decubitus prophylaxis by implementing a solution to the single-frame 2D Human Pose Estimation problem.
For this, methods of Deep Learning are employed. Two approaches are examined, a coarse-to-fine Convolutional Neural Network for direct regression of joint coordinates and a U-Net for the derivation of probability distribution heatmaps.
We conclude that training our models on a combined dataset of the publicly available Bodies at Rest and SLP data yields the best results. Furthermore, various preprocessing techniques are investigated, and a hyperparameter optimization is performed to discover an improved model architecture.
Another finding indicates that the heatmap-based approach outperforms direct regression.
This model achieves a mean per-joint position error of 9.11 cm for the Bodies at Rest data and 7.43 cm for the SLP data.
We find that it generalizes well on data from mattresses other than those seen during training but has difficulties detecting the arms correctly.
Additionally, we give a brief overview of the medical data annotation tool annoto we developed in the bachelor project and furthermore conclude that the Scrum framework and agile practices enhanced our development workflow.
The development of self-adaptive software requires the engineering of an adaptation engine that controls the underlying adaptable software by a feedback loop. State-of-the-art approaches prescribe the feedback loop in terms of numbers, how the activities (e.g., monitor, analyze, plan, and execute (MAPE)) and the knowledge are structured to a feedback loop, and the type of knowledge. Moreover, the feedback loop is usually hidden in the implementation or framework and therefore not visible in the architectural design. Additionally, an adaptation engine often employs runtime models that either represent the adaptable software or capture strategic knowledge such as reconfiguration strategies. State-of-the-art approaches do not systematically address the interplay of such runtime models, which would otherwise allow developers to freely design the entire feedback loop.
This thesis presents ExecUtable RuntimE MegAmodels (EUREMA), an integrated model-driven engineering (MDE) solution that rigorously uses models for engineering feedback loops. EUREMA provides a domain-specific modeling language to specify and an interpreter to execute feedback loops. The language allows developers to freely design a feedback loop concerning the activities and runtime models (knowledge) as well as the number of feedback loops. It further supports structuring the feedback loops in the adaptation engine that follows a layered architectural style. Thus, EUREMA makes the feedback loops explicit in the design and enables developers to reason about design decisions.
To address the interplay of runtime models, we propose the concept of a runtime megamodel, which is a runtime model that contains other runtime models as well as activities (e.g., MAPE) working on the contained models. This concept is the underlying principle of EUREMA. The resulting EUREMA (mega)models are kept alive at runtime and they are directly executed by the EUREMA interpreter to run the feedback loops. Interpretation provides the flexibility to dynamically adapt a feedback loop. In this context, EUREMA supports engineering self-adaptive software in which feedback loops run independently or in a coordinated fashion within the same layer as well as on top of each other in different layers of the adaptation engine. Moreover, we consider preliminary means to evolve self-adaptive software by providing a maintenance interface to the adaptation engine.
This thesis discusses in detail EUREMA by applying it to different scenarios such as single, multiple, and stacked feedback loops for self-repairing and self-optimizing the mRUBiS application. Moreover, it investigates the design and expressiveness of EUREMA, reports on experiments with a running system (mRUBiS) and with alternative solutions, and assesses EUREMA with respect to quality attributes such as performance and scalability.
The conducted evaluation provides evidence that EUREMA as an integrated and open MDE approach for engineering self-adaptive software seamlessly integrates the development and runtime environments using the same formalism to specify and execute feedback loops, supports the dynamic adaptation of feedback loops in layered architectures, and achieves an efficient execution of feedback loops by leveraging incrementality.
To manage tabular data files and leverage their content in a given downstream task, practitioners often design and execute complex transformation pipelines to prepare them. The complexity of such pipelines stems from different factors, including the nature of the preparation tasks, often exploratory or ad-hoc to specific datasets; the large repertory of tools, algorithms, and frameworks that practitioners need to master; and the volume, variety, and velocity of the files to be prepared. Metadata plays a fundamental role in reducing this complexity: characterizing a file assists end users in the design of data preprocessing pipelines, and furthermore paves the way for suggestion, automation, and optimization of data preparation tasks.
Previous research in the areas of data profiling, data integration, and data cleaning, has focused on extracting and characterizing metadata regarding the content of tabular data files, i.e., about the records and attributes of tables. Content metadata are useful for the latter stages of a preprocessing pipeline, e.g., error correction, duplicate detection, or value normalization, but they require a properly formed tabular input. Therefore, these metadata are not relevant for the early stages of a preparation pipeline, i.e., to correctly parse tables out of files. In this dissertation, we turn our focus to what we call the structure of a tabular data file, i.e., the set of characters within a file that do not represent data values but are required to parse and understand the content of the file. We provide three different approaches to represent file structure, an explicit representation based on context-free grammars; an implicit representation based on file-wise similarity; and a learned representation based on machine learning.
In our first contribution, we use the grammar-based representation to characterize a set of over 3000 real-world csv files and identify multiple structural issues that let files deviate from the csv standard, e.g., by having inconsistent delimiters or containing multiple tables. We leverage our learnings about real-world files and propose Pollock, a benchmark to test how well systems parse csv files that have a non-standard structure, without any previous preparation. We report on our experiments on using Pollock to evaluate the performance of 16 real-world data management systems.
Following, we characterize the structure of files implicitly, by defining a measure of structural similarity for file pairs. We design a novel algorithm to compute this measure, which is based on a graph representation of the files' content. We leverage this algorithm and propose Mondrian, a graphical system to assist users in identifying layout templates in a dataset, classes of files that have the same structure, and therefore can be prepared by applying the same preparation pipeline.
Finally, we introduce MaGRiTTE, a novel architecture that uses self-supervised learning to automatically learn structural representations of files in the form of vectorial embeddings at three different levels: cell level, row level, and file level. We experiment with the application of structural embeddings for several tasks, namely dialect detection, row classification, and data preparation efforts estimation.
Our experimental results show that structural metadata, either identified explicitly on parsing grammars, derived implicitly as file-wise similarity, or learned with the help of machine learning architectures, is fundamental to automate several tasks, to scale up preparation to large quantities of files, and to provide repeatable preparation pipelines.
Every year, the Hasso Plattner Institute (HPI) invites guests from industry and academia to a collaborative scientific workshop on the topic Operating the Cloud. Our goal is to provide a forum for the exchange of knowledge and experience between industry and academia. Co-located with the event is the HPI’s Future SOC Lab day, which offers an additional attractive and conducive environment for scientific and industry related discussions. Operating the Cloud aims to be a platform for productive interactions of innovative ideas, visions, and upcoming technologies in the field of cloud operation and administration.
In these proceedings, the results of the fifth HPI cloud symposium Operating the Cloud 2017 are published. We thank the authors for exciting presentations and insights into their current work and research. Moreover, we look forward to more interesting submissions for the upcoming symposium in 2018.