Institut für Informatik und Computational Science
Refine
Year of publication
Document Type
- Article (576)
- Doctoral Thesis (203)
- Monograph/Edited Volume (135)
- Other (28)
- Conference Proceeding (17)
- Part of a Book (12)
- Master's Thesis (10)
- Postprint (10)
- Preprint (4)
- Bachelor Thesis (1)
Is part of the Bibliography
- yes (998) (remove)
Keywords
- answer set programming (13)
- Answer Set Programming (10)
- Answer set programming (10)
- Machine Learning (7)
- Maschinelles Lernen (7)
- Antwortmengenprogrammierung (6)
- E-Learning (6)
- Informatik (6)
- Modellierung (5)
- Informatikdidaktik (4)
- Internet of Things (4)
- MQTT (4)
- higher education (4)
- machine learning (4)
- security (4)
- Algorithms (3)
- Digitale Medien (3)
- EEG (3)
- Equilibrium logic (3)
- Komplexität (3)
- Machine learning (3)
- Modeling (3)
- Ontologie (3)
- Optimization (3)
- Semantic Web (3)
- Softwareentwicklung (3)
- didactics (3)
- education (3)
- formal languages (3)
- image processing (3)
- monitoring (3)
- teacher training (3)
- verification (3)
- 3D visualization (2)
- ASIC (2)
- Adaptivity (2)
- Algorithmen (2)
- Analytical models (2)
- Assessment (2)
- Autismus (2)
- Automata systems (2)
- Bildungstechnologien (2)
- Bildverarbeitung (2)
- Code (2)
- Codierungstheorie (2)
- Complexity (2)
- Computergrafik (2)
- Computersicherheit (2)
- Coq (2)
- Debugging (2)
- Deep Learning (2)
- Didaktik (2)
- Digitalisierung (2)
- E-learning (2)
- Educational Technologies (2)
- Event mapping (2)
- FMC (2)
- FPGA (2)
- Fault tolerance (2)
- Fehlererkennung (2)
- Hochschuldidaktik (2)
- Hochschullehre (2)
- ICA (2)
- IT-Infrastruktur (2)
- Informatikstudium (2)
- Internet (2)
- Knowledge Representation and Reasoning (2)
- Konstruktivismus (2)
- Künstliche Intelligenz (2)
- Lernumgebung (2)
- Lindenmayer systems (2)
- Mensch-Technik-Interaktion (2)
- Methodik (2)
- Middleware (2)
- Modell (2)
- Modelling (2)
- Non-monotonic reasoning (2)
- Onlinelehre (2)
- Ontology (2)
- Optimierung (2)
- Parameterized complexity (2)
- Preference Handling (2)
- Process (2)
- Process mining (2)
- Prototyping (2)
- Prozess (2)
- Prozessmodellierung (2)
- ResNet (2)
- Software Engineering (2)
- Synthese (2)
- Systemstruktur (2)
- TPACK (2)
- Texturen (2)
- Theory (2)
- Tracking (2)
- Treewidth (2)
- User Experience (2)
- Visualisierung (2)
- Vorhersage (2)
- anti-cancer drugs (2)
- argument mining (2)
- automatic feedback (2)
- bioinformatics (2)
- code (2)
- complexity (2)
- computer graphics (2)
- computer science education (2)
- concurrent checking (2)
- cooperating systems (2)
- deep neural networks (2)
- drug-sensitivity prediction (2)
- e-learning (2)
- edge computing (2)
- embedded systems (2)
- face tracking (2)
- facial expression (2)
- firmware update (2)
- geovisualization (2)
- human computer interaction (2)
- informatics (2)
- knowledge representation and nonmonotonic reasoning (2)
- lesson planning (2)
- lesson preparation (2)
- logic programming (2)
- maschinelles Lernen (2)
- model (2)
- non-photorealistic rendering (2)
- online learning (2)
- parsing (2)
- perception (2)
- perception differences (2)
- physical computing (2)
- policy evaluation (2)
- radhard design (2)
- reliability (2)
- scientific workflows (2)
- self-adaptive multiprocessing system (2)
- single event upset (2)
- software development (2)
- solar particle event (2)
- support system (2)
- systems biology (2)
- test (2)
- textures (2)
- virtual 3D city models (2)
- virtual mobility (2)
- virtuelle 3D-Stadtmodelle (2)
- visualization (2)
- 'Peer To Peer' (1)
- (FPGA) (1)
- (SET) count rate (1)
- 13C metabolic flux analysis (1)
- 2-tag system (1)
- 3D Computer Grafik (1)
- 3D Computer Graphics (1)
- 3D Drucken (1)
- 3D Linsen (1)
- 3D Semiotik (1)
- 3D Visualisierung (1)
- 3D computer graphics (1)
- 3D lenses (1)
- 3D modeling (1)
- 3D printing (1)
- 3D semiotics (1)
- 3D-Stadtmodelle (1)
- 3d city models (1)
- 6LoWPAN (1)
- ADFS (1)
- AODV (1)
- ASIC (Applikationsspezifische Integrierte Schaltkreise) (1)
- ASP (Answer Set Programming) (1)
- Abbrecherquote (1)
- Absorbed dose (1)
- Abstraction (1)
- Abstraktion (1)
- Accepting Grammars (1)
- Access control (1)
- Ackerschmalwand (1)
- Active Directory Federation Services (1)
- Active Evaluation (1)
- Active evaluation (1)
- Ad hoc routing (1)
- Adaptivität (1)
- Advanced Video Codec (AVC) (1)
- Adversarial Learning (1)
- Aggregates (1)
- Aktive Evaluierung (1)
- Akzeptierende Grammatiken (1)
- Algorithm configuration (1)
- Algorithm portfolios (1)
- Algorithmenablaufplanung (1)
- Algorithmenkonfiguration (1)
- Algorithmenselektion (1)
- Alignment (1)
- Android Security (1)
- Android hybrid apps (1)
- Angewandte Spieltheorie (1)
- Animal building (1)
- Anisotroper Kuwahara Filter (1)
- Anleitung (1)
- Answer Set Solving modulo Theories (1)
- Anti-cancer drugs (1)
- Antwortmengen Programmierung (1)
- App (1)
- Application Aggregation (1)
- Applications and Software Development (1)
- Applied Game Theory (1)
- Apps (1)
- Arabidopsis thaliana (1)
- Argumentation (1)
- Argumentation structure (1)
- Artificial Intelligence (1)
- Artificial Neuronal Network (1)
- Asynchrone Schaltung (1)
- Attention (1)
- Aufmerksamkeit (1)
- Aufzählung (1)
- Augenbewegungen (1)
- Augmentation (1)
- Augmented and virtual reality (1)
- Ausbildung (1)
- Ausreissererkennung (1)
- Authorization (1)
- Autism (1)
- Automated parallelization (1)
- Automatic UI Generation (1)
- Automatically controlled windows (1)
- Autorisierung (1)
- BCH (1)
- BCH code (1)
- BCH-Code (1)
- BCI (1)
- BSS (1)
- Bachelorstudierende der Informatik (1)
- Backdoors (1)
- Barcode (1)
- Batch processing (1)
- Baumweite (1)
- Bean (1)
- Behavior (1)
- Benchmark testing; (1)
- Benutzeroberfläche (1)
- Benutzungsschnittstellen Ontologien (1)
- Berliner Modell (1)
- Berührungseingaben (1)
- Betrachtungsebenen (1)
- Beweis (1)
- Beweisassistent (1)
- Beweistheorie (1)
- Beweisumgebung (1)
- Bilddatenanalyse (1)
- Bildung (1)
- Binäres Entscheidungsdiagramm (1)
- Bio-jETI (1)
- Biocomputing (1)
- Bioelektrisches Signal (1)
- Bioinformatik (1)
- Blind users (1)
- Boolean constraint solver (1)
- Boolean logic models (1)
- Boosting (1)
- Bot Detection (1)
- Brain Computer Interface (1)
- Business Process (1)
- Business Process Models (1)
- Business process intelligence (1)
- CASP (Constraint Answer Set Programming) (1)
- CP-Logic (1)
- CSC (1)
- Cactus (1)
- Campus (1)
- CertiCoq (1)
- Chipkarten (1)
- Choreographien (1)
- Circuit faults (1)
- CityGML (1)
- Classification (1)
- Clock tree (1)
- Cloud (1)
- Cluster Computing (1)
- Cluster computing (1)
- Clusteranalyse (1)
- Code generation (1)
- Codierung (1)
- Coding theory (1)
- Cognitive Apprenticeship (1)
- Coherent phonons (1)
- Combinatorial multi-objective optimization (1)
- Common Spatial Pattern (1)
- Complementary Circuits (1)
- Complex optimization (1)
- Compliance (1)
- Composed UIs (1)
- Composition (1)
- Computational Complexity (1)
- Computational Science (1)
- Computational complexity (1)
- Computational grid (1)
- Computer Science (1)
- Computer Science Education (1)
- Computer security (1)
- Computergestützes Training (1)
- Computing with DNA (1)
- Conceptual (1)
- Conformant Planning (1)
- Conrad Hal Waddington (1)
- Constraint (1)
- Constraint Solving (1)
- Constraint satisfaction (1)
- Constraint-Programmierung (1)
- Constructive solid geometry (1)
- Context awareness (1)
- Contextualized learning (1)
- Continuous Testing (1)
- Continuous Versioning (1)
- Controlled Derivations (1)
- Convolution (1)
- Course timetabling (1)
- Covariate Shift (1)
- Curry (1)
- Customer ownership (1)
- D-galactosamine (1)
- DDoS (1)
- DMR (1)
- DNA (1)
- DNA computing (1)
- DNA hairpin formation (1)
- DNS (1)
- DPLL (1)
- DRMAA (1)
- DRMS (1)
- Data federation (1)
- Database (1)
- Deal of the Day (1)
- Decidability (1)
- Declarative Problem Solving (1)
- Declare (1)
- Deduction (1)
- Deep learning (1)
- Defining characteristics of physical computing (1)
- Dempster-Shafer-Theorie (1)
- Dempster–Shafer theory (1)
- Denotational semantics (1)
- Description Logics (1)
- Design (1)
- Design for testability (DFT) (1)
- Deskriptive Logik (1)
- Deutschland (1)
- Diagonalisierung (1)
- Dialog-based User Interfaces (1)
- Dialogbasierte Benutzerschnittstellen (1)
- Didaktik der Informatik (1)
- Didaktische Konzepte (1)
- Dienst-Ökosysteme (1)
- Dienstkomposition (1)
- Dienstplattform (1)
- Differenz von Gauss Filtern (1)
- Digital Design (1)
- Digital Game Based Learning (1)
- Digital Media (1)
- Digital image analysis (1)
- Digitalisation (1)
- Digitalization (1)
- Distributed Computing (1)
- Diversität (1)
- Domain-Specific Languages (1)
- Domänenspezifische Sprachen (1)
- Dose rate (1)
- Double cell upsets (DCUs) (1)
- Dreidimensionale Computergraphik (1)
- Dynamic Programming (1)
- Dynamical X-ray theory (1)
- Dynamische Programmierung (1)
- Dynamische Rekonfiguration (1)
- E-Assessment (1)
- E-Government (1)
- E-Klausuren (1)
- E-Portfolio (1)
- E-teaching (1)
- EDC (1)
- Echtzeitanwendung (1)
- Edge Computing (1)
- Educational game (1)
- Educational timetabling (1)
- Eingabegenauigkeit (1)
- Eisenbahnnetz (1)
- Elektroencephalographie (1)
- Elektronisches Prüfen (1)
- Emotionen (1)
- Emotionsforschung (1)
- Encoding (1)
- Engines (1)
- Enterprise Architecture (1)
- Enterprise Search (1)
- Entity Linking (1)
- Entscheidungsbäume (1)
- Entwurf (1)
- Entwurfsmuster (1)
- Entwurfsmuster für SOA-Sicherheit (1)
- Entwurfsprinzipien (1)
- Entwurfsraumexploration (1)
- Epigenetic landscape (1)
- Epistemic Logic Programs (1)
- Erfüllbarkeit einer Formel der Aussagenlogik (1)
- Erfüllbarkeitsproblem (1)
- Erklärbarkeit (1)
- Error Estimation (1)
- Error-Detection Circuits (1)
- European Bioinformatics Institute (1)
- Evaluation (1)
- Evaluierung semantischer Suchmaschinen (1)
- Evidenztheorie (1)
- Evolution (1)
- Experimentation (1)
- Explainability (1)
- Explicit negation (1)
- Exploration (1)
- Explore-first Programming (1)
- Exponential Time Hypothesis (1)
- Exponentialzeit Hypothese (1)
- Extensibility (1)
- Extreme Model-Driven Development (1)
- FEDC (1)
- FMC-QE (1)
- FOSS (1)
- Fault Localization (1)
- Fault tolerant systems (1)
- Feature Combination (1)
- Feature extraction (1)
- Feedback (1)
- Fehlende Daten (1)
- Fehlerkorrektur (1)
- Fehlerschätzung (1)
- Fehlvorstellung (1)
- Field programmable gate arrays (1)
- Finite automata (1)
- Flip-flops (1)
- Flussgesteuerter Bilateraler Filter (1)
- Focus+Context Visualization (1)
- Fokus-&-Kontext Visualisierung (1)
- Forgetting (1)
- Formalismus (1)
- Formalitätsgrad (1)
- Formeln der quantifizierten Aussagenlogik (1)
- Forschendes Lernen (1)
- Forschungsdatenmanagement (1)
- Framework (1)
- Freshmen (1)
- GERBIL (1)
- GIS-Dienstkomposition (1)
- GPU (1)
- Game-Design-Elemente (1)
- Game-based learning (1)
- Gamification (1)
- Gebrauchstauglichkeit (1)
- Gebäudemodelle (1)
- Gehirn-Computer-Schnittstelle (1)
- Geländemodelle (1)
- Generalisierung (1)
- Generative Programmierung (1)
- Generative Programming (1)
- Geodaten (1)
- Geometrieerzeugung (1)
- Geovisualisierung (1)
- Geräte-Treiber (1)
- Geschäftsprozess (1)
- Geschäftsprozessmodelle (1)
- Gesichtsausdruck (1)
- Gesteuerte Ableitungen (1)
- Gesture input (1)
- Gleichheit (1)
- Globus (1)
- Grammar Systems (1)
- Grammatikalische Inferenz (1)
- Grammatiksysteme (1)
- Graph Convolutional Neural Networks (1)
- Graph Embedding (1)
- Graph-basiertes Ranking (1)
- Graphfärbung (1)
- Green computing (1)
- Grid (1)
- Grid Computing (1)
- Grounded theory (1)
- Grounding Theory (1)
- H.264 (1)
- HCI (1)
- HDI (1)
- Hairpin completions (1)
- Hairpin reductions (1)
- Hardware Design (1)
- Hardware accelerator (1)
- Hardware-Software-Co-Design (1)
- Hauptkomponentenanalyse (1)
- Heat diffusion (1)
- Heterogenität (1)
- Hierarchically configurable mask register (1)
- High-Level Synthesis (1)
- Histograms (1)
- Hochschul-Apps (1)
- Hochschul-Cloud (1)
- Hochschulbildung (1)
- Hochschulsystem (1)
- Https traffic (1)
- Human Factors (1)
- Human-Technology Interaction (1)
- Hurricane Sandy (1)
- Hybrid App (1)
- I/O-effiziente Algorithmen (1)
- ICT (1)
- IP core (1)
- IT security (1)
- IT-Security (1)
- IT-Sicherheit (1)
- IaaS (1)
- Identifiers (1)
- Image and video stylization (1)
- Image resolution (1)
- Imperative calculi (1)
- Improving classroom (1)
- Incoherent phonons (1)
- Incremental answer set programming (1)
- Industrie 4.0 (1)
- Industry 4.0 (1)
- Inference (1)
- Informatics (1)
- Informatik-Studiengänge (1)
- Informatiksystem (1)
- Informatikunterricht (1)
- Informatikvoraussetzungen (1)
- Information Transfer Rate (1)
- Information federation (1)
- Information integration (1)
- Information retrieval (1)
- Information security (1)
- Informationsextraktion (1)
- Infrastruktur (1)
- Inkonsistenz (1)
- Innovation (1)
- Inquiry-based learning (1)
- Insurance industry (1)
- Integrated circuit modeling (1)
- Integration (1)
- Interactive Rendering (1)
- Interactive system (1)
- Interaktionsmodel (1)
- Interaktionsmodellierung (1)
- Interaktionstechniken (1)
- Interaktives Rendering (1)
- Interaktives System (1)
- Interface design (1)
- Internet Security (1)
- Internet-Sicherheit (1)
- Interoperability (1)
- Interoperabilität (1)
- Interpretability (1)
- Interpretierbarkeit (1)
- Intuition (1)
- IoT (1)
- Job monitoring (1)
- Job submission (1)
- Kartografisches Design (1)
- Kern-PCA (1)
- Kernel (1)
- Kernelization (1)
- Kernmethoden (1)
- Key Competencies (1)
- Key input (1)
- Klassifikation (1)
- Klassifikation mit großem Margin (1)
- Klassifikator-Kalibrierung (1)
- Klimafolgenanalyse (1)
- Klimawandel (1)
- Knowledge (1)
- Knowledge Management (1)
- Knowledge representation (1)
- Kommunikation (1)
- Kommunismus (1)
- Kompetenz (1)
- Kompilation (1)
- Komplexitätsbewältigung (1)
- Komplexitätstheorie (1)
- Komposition (1)
- Konnektionskalkül (1)
- Konzeptionell (1)
- Kybernetik (1)
- Künstliche Neuronale Netzwerke (1)
- L systems (1)
- LBA problem (1)
- LDPC code (1)
- LDPC-Code (1)
- Landmark visibility (1)
- Landmarken (1)
- Large Margin Classification (1)
- Laser Cutten (1)
- Learning (1)
- Learning Analytics (1)
- Learning environment (1)
- Lebenslanges Lernen (1)
- Leftmost Derivations (1)
- Lehre (1)
- Lehrer (1)
- Lehrkräfteausbildung (1)
- Leistungsvorhersage (1)
- Lernen (1)
- Lernsoftware (1)
- Life-Long Learning (1)
- Linked Data Anwendungen (1)
- Linked Data Application Modelling (1)
- Linksableitungen (1)
- Linux (1)
- Linux device drivers (1)
- Literature mining (1)
- Liver neoplasms (1)
- Load Balancing (1)
- Localization (1)
- Location awareness (1)
- Logic Programming (1)
- Logic programming (1)
- Logics (1)
- Logik (1)
- Logiksynthese (1)
- Loss (1)
- Low Latency (1)
- Lower Bounds (1)
- Loyalty (1)
- M2M (1)
- MEG (1)
- MFA (1)
- Magnetoencephalographie (1)
- Malware (1)
- Markov processes (1)
- Masking of X-values (1)
- Massenklausuren (1)
- Mathematical Optimization (1)
- Mathematikdidaktik (1)
- Mathematikphilosophie (1)
- Mathematische Optimierung (1)
- Matrizen-Eigenwertaufgabe (1)
- Media in education (1)
- Megamodel (1)
- Megamodell (1)
- Mehrklassen-Klassifikation (1)
- Mensch-Computer-Interaktion (1)
- Message Passing Interface (1)
- Meta-Programming (1)
- Metamodell (1)
- Methoden der semantischen Suche (1)
- Methodology (1)
- Metric learning (1)
- Migration (1)
- Minimal perturbation problems (1)
- Mischmodelle (1)
- Mischung <Signalverarbeitung> (1)
- Mobile App (1)
- Mobile Campus Application (1)
- Mobile Learning (1)
- Mobile application (1)
- Mobile devices (1)
- Mobile learning (1)
- Mobiles Lernen (1)
- Mobilgeräte (1)
- Model Based Engineering (1)
- Model Checking (1)
- Model Driven Architecture (1)
- Model Driven UI Development (1)
- Model Management (1)
- Model checking (1)
- Model-Driven Engineering (1)
- Model-Driven Software Development (1)
- Modell Management (1)
- Modell-driven Security (1)
- Modell-getriebene Sicherheit (1)
- Modellbasiert (1)
- Modellgetriebene Architektur (1)
- Modellgetriebene Entwicklung (1)
- Modellgetriebene Softwareentwicklung (1)
- Modellgetriebene UI Entwicklung (1)
- Molekulare Bioinformatik (1)
- Motivation (1)
- Multi Task Learning (1)
- Multi-Class (1)
- Multi-Task-Lernen (1)
- Multi-objective optimization (1)
- Multi-sided platforms (1)
- Multimedia (1)
- Multimodal User Interfaces (1)
- Multimodale Benutzerschnittstellen (1)
- Multiple interpretation scheme (1)
- Multiprocessor (1)
- Multiprozessor (1)
- N-temperature model (1)
- NETCONF (1)
- NFC (1)
- NUI (1)
- Nash equilibrium (1)
- Natural language processing (1)
- Natural ventilation (1)
- Navigation (1)
- Network (1)
- Network Management (1)
- Network security (1)
- Netzwerk Management (1)
- Netzwerke (1)
- Neural networks (1)
- Neuronales Netz (1)
- New On-Line Error-Detection Methode (1)
- Next Generation Network (1)
- Nicht-photorealistisches Rendering (1)
- Nichtfotorealistische Bildsynthese (1)
- Non-Monotonic (1)
- Nonmonotonic reasoning (1)
- Nutzungserlebnis (1)
- Nutzungsinteresse (1)
- OBI (1)
- OCCI (1)
- OSSE (1)
- Objektive Schwierigkeit (1)
- Omega (1)
- Ontologien (1)
- Ontologies (1)
- Open Badge Infrastructure (1)
- Open Badges (1)
- Open Source (1)
- Operation problem (1)
- Optimierungsproblem (1)
- Parallel Programming (1)
- Parallel SAT solving (1)
- Parallel job execution time estimation (1)
- Parallele Datenverarbeitung (1)
- Paralleles Rechnen (1)
- Parallelrechner (1)
- Parameterized Complexity (1)
- Parametrisierte Komplexität (1)
- Parsing (1)
- Particle detector (1)
- Partizipation (1)
- Patterns (1)
- Pedagogical issues (1)
- Pedestrian navigation (1)
- Peer-Review (1)
- Peer-to-Peer-Netz ; GRID computing ; Zuverlässigkeit ; Web Services ; Betriebsmittelverwaltung ; Migration (1)
- Performance Evaluation (1)
- Performance Prediction (1)
- Personal Learning Environment (1)
- Personalization (1)
- Persönliche Lernumgebung (1)
- Pervasive computing (1)
- Pervasive game (1)
- Pervasive learning (1)
- Phantoms (1)
- Planar tactile display (1)
- Planing (1)
- Plant identification (1)
- Platzierung (1)
- Polarization (1)
- Policy Enforcement (1)
- Power Monitoring (1)
- Pre-RS Traceability (1)
- Prediction Game (1)
- Predictive Models (1)
- Privacy Protection (1)
- Probleme in der Studie (1)
- Problemlösen (1)
- Process Management (1)
- Process model analysis (1)
- Process modeling (1)
- Product lifecycle management (1)
- Professionalisierung (1)
- Professoren (1)
- Programmierung (1)
- Programming (1)
- Programming by optimization (1)
- Proof Theory (1)
- Prototyp (1)
- Prozess Verbesserung (1)
- Prozesse (1)
- Prozessmanagement (1)
- Prozessmodell (1)
- Prozesssynchronisierung (1)
- Prädiktionsspiel (1)
- Präferenzen (1)
- Prüfungsoptimierung (1)
- Quantified Boolean Formula (QBF) (1)
- Quantitative Modeling (1)
- Quantitative Modellierung (1)
- Queuing Theory (1)
- RADFET (1)
- RADFETs (1)
- REST (1)
- RSA triangle (1)
- Radiation hardness (1)
- Random access memory (1)
- Ranking (1)
- Reasoning (1)
- Reconfigurable (1)
- Reflexion (1)
- Region of Interest (1)
- Regression (1)
- Regularisierung (1)
- Regularization (1)
- Rekonfiguration (1)
- Rendering (1)
- Reparatur (1)
- Reproducibility of results (1)
- Reuseable UIs (1)
- Reversibility (1)
- SAMR (1)
- SET pulsewidth distribution (1)
- SMT (SAT Modulo Theories) (1)
- SOA (1)
- SOA Security Pattern (1)
- SSO (1)
- STG decomposition (1)
- STG-Dekomposition (1)
- SWOT (1)
- SaaSAbstract (1)
- Sample Selection Bias (1)
- Satisfiability (1)
- Scalability (1)
- Scale-invariant feature transform (SIFT) (1)
- Scene graph systems (1)
- Schlüsselkompetenzen (1)
- Schulmaterial (1)
- Scientific images (1)
- Screen reader (1)
- Seamless learning (1)
- Search problems (1)
- Security (1)
- Security Modelling (1)
- Segmentierung (1)
- Selektionsbias (1)
- Self-Checking Circuits (1)
- Self-adaptive MPSoC (1)
- Semantic Interoperability (1)
- Semantic Search (1)
- Semantic data (1)
- Semantic web (1)
- Semantik Web (1)
- Semantische Suche (1)
- Semilinearity property (1)
- Sensornetzwerke (1)
- Sequence embeddings (1)
- Service Creation (1)
- Service Delivery Platform (1)
- Service Ecosystems (1)
- Service Oriented Architectures (1)
- Service convergence (1)
- Service orientation (1)
- Service-Orientierte Architekturen (1)
- Service-oriented Architecture (1)
- Service-oriented Architectures (1)
- Service-oriented architecture (1)
- Serviceorientierte Architektur (1)
- Shader (1)
- Sharing (1)
- Sicherheitsmodellierung (1)
- Signal Processing (1)
- Signal processing (1)
- Signaling transduction networks (1)
- Signalquellentrennung (1)
- Signaltrennung (1)
- Simulation (1)
- Simulations (1)
- Simultane Diagonalisierung (1)
- Single Event Transient (1)
- Single Sign On (1)
- Single Trial Analysis (1)
- Single event effect (1)
- Single event upsets (1)
- Single-event transient (SET) (1)
- Skelettberechnung (1)
- Social Media Analysis (1)
- Software (1)
- Software architecture (1)
- Software-basierte Cache-Kohärenz (1)
- Softwarearchitektur (1)
- Sonnenteilchen-Ereignis (1)
- Spam (1)
- Spam Filtering (1)
- Spam-Erkennung (1)
- Spam-Filter (1)
- Spam-Filtering (1)
- Spatio-Spectral Filter (1)
- Spawning (1)
- Speicher (1)
- Spielbasiertes Lernen (1)
- Splicing (1)
- Splicing processor (1)
- Sprachdesign (1)
- Static Analysis (1)
- Statistical Tests (1)
- Statistical relational learning (1)
- Statistikprogramm R (1)
- Statistische Tests (1)
- Stilisierung (1)
- Stochastic relational process (1)
- Strahlungshartes Design (1)
- Strahlungshärte Entwurf (1)
- Strategie (1)
- Stromverbrauchüberwachung (1)
- Strong equivalence (1)
- Structural equation modeling (1)
- Structuring (1)
- Strukturierung (1)
- Studentenerwartungen (1)
- Studentenhaltungen (1)
- Studentenjobs (1)
- Studienabbrecher (1)
- Studiendauer (1)
- Studieneinstieg (1)
- Studienorganisation (1)
- Suche (1)
- Support Vectors (1)
- Support-Vector Lernen (1)
- System Biologie (1)
- System structure (1)
- Systembiologie (1)
- Systementwurf (1)
- Systems biology (1)
- Systems of parallel communicating (1)
- Szenengraph (1)
- TMR (1)
- TPTP (1)
- Tailored UI Variants (1)
- Taktik (1)
- Teilnehmerzertifikate (1)
- Telekommunikation (1)
- Temporal Answer Set Solving (1)
- Temporal Logic (1)
- Temporallogik (1)
- Temporäre Anbindung (1)
- Terminologische Logik (1)
- Test (1)
- Theoretische Informatik (1)
- Theoretischen Vorlesungen (1)
- Theory formation (1)
- Thermoelasticity (1)
- Time Augmented Petri Nets (1)
- Time Series Analysis (1)
- Time series (1)
- Tomography (1)
- Tool (1)
- Tools (1)
- Traceability (1)
- Traffic data (1)
- Transformation (1)
- Tree decomposition (1)
- Treewidth-aware reductions (1)
- Triple modular redundancy (TMR) (1)
- Tumor types (1)
- Turing machine (1)
- Type and effect systems (1)
- UAV imagery (1)
- UI Components (1)
- UI Metamodels (1)
- UI-Komponenten (1)
- UX (1)
- Ubiquitous learning (1)
- Ultrafast dynamics (1)
- Unabhängige Komponentenanalyse (1)
- Unary languages (1)
- Uniform Access Principle (1)
- University Service Bus (1)
- Universität Bagdad (1)
- Universität Potsdam (1)
- Universitätseinstellungen (1)
- Untere Schranken (1)
- Unterrichtswerkzeuge (1)
- Unvollständigkeit (1)
- Usability (1)
- Usability testing (1)
- Usage Interest (1)
- User Interface Ontologies (1)
- User Interfaces (1)
- User submission pattern (1)
- User-centred design (1)
- VGG16 (1)
- VM (1)
- Value network (1)
- Verhalten (1)
- Verification (1)
- Verifikation (1)
- Verletzung Auflösung (1)
- Verletzung Erklärung (1)
- Verteiltes Rechnen (1)
- Verteilungsunterschied (1)
- Violation Explanation (1)
- Violation Resolution (1)
- Virtual Reality (1)
- Virtual reality (1)
- Virtuelles 3D Stadtmodell (1)
- Visual metaphor (1)
- Vorhersagemodelle (1)
- Wahrnehmung (1)
- Wahrnehmung von Arousal (1)
- Wahrnehmungsunterschiede (1)
- Warteschlangentheorie (1)
- Web Services (1)
- Web Sites (1)
- Web of Data (1)
- Webanwendung (1)
- Webseite (1)
- Well-structuredness (1)
- Wetterextreme (1)
- Wireless Sensor Networks (1)
- Wirtschaftsinformatik (1)
- Wissen (1)
- Wissenschaftlichesworkflows (1)
- Wissensmanagement (1)
- Wissensrepräsentation und -verarbeitung (1)
- Wissensrepräsentation und Schlussfolgerung (1)
- Wohlstrukturiertheit (1)
- Word embeddings (1)
- Workflow (1)
- X-masking (1)
- X-ray computed (1)
- X-values (1)
- ZQSA (1)
- ZQSAT (1)
- Zeitbehaftete Petri Netze (1)
- Zero-Suppressed Binary Decision Diagram (ZDD) (1)
- Zuverlässigkeitsanalyse (1)
- abstraction (1)
- accepting grammars (1)
- action and change (1)
- activities (1)
- activity (1)
- acute liver failure (1)
- acyclicity properties (1)
- adaptiv (1)
- adaptive (1)
- adversarial classification (1)
- algorithm configuration (1)
- algorithm schedules (1)
- algorithm scheduling (1)
- algorithm selection (1)
- algorithms (1)
- analysis (1)
- animated PCA (1)
- animierte PCA (1)
- anisotropic Kuwahara filter (1)
- annealing (1)
- anxiety (1)
- approximate joint diagonalization (1)
- approximate model counting (1)
- architecture (1)
- argumentation (1)
- argumentation structure (1)
- arithmethische Prozeduren (1)
- arithmetic procedures (1)
- arousal (1)
- arousal perception (1)
- artificial intelligence (1)
- artistic rendering (1)
- asynchronous circuit (1)
- asynchrounous design (1)
- authentication (1)
- autism (1)
- automata (1)
- automated driving (1)
- automated guided vehicle routing (1)
- automated planning (1)
- automatic theorem prover (1)
- automatisierter Theorembeweiser (1)
- behavioral (1)
- behavioral abstraction (1)
- belief merging (1)
- belief revision (1)
- benchmark (1)
- bibliometric analysis (1)
- bild (1)
- bio-computing (1)
- biometrics (1)
- biometrische Identifikation (1)
- blind source separation (1)
- block representation (1)
- bootstrapping (1)
- brain-computer interface (1)
- building models (1)
- bundled data (1)
- business informatics (1)
- camera sensor (1)
- car assembly operations (1)
- cartographic design (1)
- cellular automata (1)
- changing the study field (1)
- changing the university (1)
- choreographies (1)
- circuit Faults (1)
- citation analysis (1)
- classifier calibration (1)
- classroom material (1)
- click controller (1)
- climate change (1)
- climate impact analysis (1)
- clocks (1)
- clustering (1)
- co-citation analysis (1)
- co-occurrence analysis (1)
- code generation (1)
- cognitive apprenticeship (1)
- coherence relation (1)
- coherence-enhancing filtering (1)
- collaborative learning (1)
- combinatorial optimization problems (1)
- combined task and motion planning (1)
- common spatial patterns (1)
- communication (1)
- competition (1)
- compilation (1)
- complex networks (1)
- compliance (1)
- computational biology (1)
- computational methods (1)
- computational thinking (1)
- computer security (1)
- computer vision (1)
- computergestützte Methoden (1)
- concession (1)
- conductive argument (1)
- connection calculus (1)
- connective (1)
- connectivity (1)
- consistency (1)
- consistency checking (1)
- consistency measures (1)
- constraint (1)
- constraint programming (1)
- constraints (1)
- constructivism (1)
- construktivism (1)
- context-free grammar (1)
- context-sensitive (1)
- continuous (1)
- contrast (1)
- controlled vocabularies (1)
- corpus analysis (1)
- correlated errors (1)
- course timetabling (1)
- craters (1)
- crop (1)
- debugging (1)
- decidability questions (1)
- decision trees (1)
- declarative problem solving (1)
- deep learning (1)
- deep residual networks (1)
- degree of formality (1)
- degree of non-context-freeness (1)
- degree of non-regularity (1)
- degree of non-regulation (1)
- depression (1)
- design (1)
- design flow (1)
- design principles (1)
- design space exploration (1)
- determinism (1)
- detrending (1)
- developmental systems (1)
- diagnosis (1)
- didaktische Rekonstruktion (1)
- difference of Gaussians (1)
- digital circuit (1)
- digital design (1)
- digitale Hochschullehre (1)
- digitally-enabled pedagogies (1)
- domain-specific APIs (1)
- dropout (1)
- drug discovery (1)
- dynamic (1)
- dynamic classification (1)
- dynamic reconfiguration (1)
- dynamic service binding (1)
- dynamisch (1)
- dynamische Klassifikation (1)
- e-Learning (1)
- eGovernment (1)
- eLectures (1)
- economic ripples (1)
- educational reconstruction (1)
- educational systems (1)
- educational timetabling (1)
- eingebettete Systeme (1)
- einseitige Kommunikation (1)
- email spam detection (1)
- emission factor (1)
- emotion (1)
- emotion representation (1)
- emotion research (1)
- endothelin (1)
- endothelin-converting enzyme (1)
- ensemble kalman filter (1)
- ensemble methods (1)
- enterprise search (1)
- entity alignment (1)
- enumeration (1)
- epistemic logic programs (1)
- epistemic specifications (1)
- equality (1)
- error correction (1)
- error detection (1)
- error propagation (1)
- evaluation (1)
- event-related desynchronization (1)
- evidence theory (1)
- evolution (1)
- explicit negation (1)
- external ambiguity (1)
- external memory algorithms (1)
- extreme weather (1)
- eye movements (1)
- fading (1)
- fault tolerance (1)
- field-programmable gate array (1)
- finite model computation (1)
- finite state sequential transducers (1)
- flow-based bilateral filter (1)
- formal (1)
- formal argumentation systems (1)
- formalism (1)
- freie Daten (1)
- freie Software (1)
- functions (1)
- gait (1)
- game based learning (1)
- game design elements (1)
- game-based learning (1)
- gap-filling (1)
- generalization (1)
- geometry generation (1)
- geospatial data (1)
- geospatial services (1)
- global constraints (1)
- globale Constraints (1)
- gradient boosting (1)
- grammar (1)
- grammar inference (1)
- graph analysis (1)
- graph clustering (1)
- graph-based ranking (1)
- greenhouse gas (1)
- hardware accelerator (1)
- hardware architecture (1)
- hardware design (1)
- hardware-software-codesign (1)
- high-throughput analysis (1)
- human-technology interaction (1)
- hybrid (1)
- hybrid semantic search (1)
- hybrid solving (1)
- hybride semantische Suche (1)
- hybrides Problemlösen (1)
- ice harboring (1)
- image (1)
- image classification (1)
- image data analysis (1)
- image recognition (1)
- imaging (1)
- impacts (1)
- incompleteness (1)
- inconsistency (1)
- incremental SVM (1)
- independent component analysis (1)
- indirect economic impacts (1)
- indirekte ökonomische Effekte (1)
- informal and formal learning (1)
- informal logic (1)
- information extraction (1)
- information flow control (1)
- information retrieval (1)
- informatische Bildung im Sekundarbereich (1)
- infrastructure (1)
- input accuracy (1)
- interaction modeling (1)
- interaction techniques (1)
- internal ambiguity (1)
- intrusion detection (1)
- intuition (1)
- irradiation (1)
- joint lab (1)
- kernel PCA (1)
- kernel methods (1)
- key competences in physical computing (1)
- kidney cancer (1)
- knowledge representation and reasoning (1)
- konvergente Dienste (1)
- landmarks (1)
- language design (1)
- latches (1)
- lautes Denken (1)
- leanCoP (1)
- learning environment (1)
- lebenslanges Lernen (1)
- leftmost derivations (1)
- linear code (1)
- linear programming (1)
- linearer Code (1)
- locomotion (1)
- logic (1)
- logic programming methodology and applications (1)
- logic synthesis (1)
- logic-based modeling (1)
- logical errors (1)
- logical signaling networks (1)
- logische Ergänzung (1)
- logische Fehler (1)
- logische Programmierung (1)
- logische Signalnetzwerke (1)
- loop formulas (1)
- loose programming (1)
- loss propagation (1)
- lunar exploration (1)
- machine learning algorithms (1)
- macro-economic modelling (1)
- makroökonomische Modellierung (1)
- malware detection (1)
- manipulation planning (1)
- map/reduce (1)
- maschninelles Lernen (1)
- mathematics education (1)
- measure development (1)
- media (1)
- medical (1)
- medizinisch (1)
- meta model (1)
- metabolic network (1)
- metabolism (1)
- metabolomics (1)
- metadata (1)
- metastasis (1)
- methodology (1)
- middleware (1)
- misconception (1)
- mixture models (1)
- mobile Applikationen (1)
- mobile devices (1)
- mobile learning (1)
- mobile technologies and apps (1)
- mobiles lernen (1)
- model-based (1)
- model-driven architecture (1)
- modeling (1)
- molecular networks (1)
- molekulare Netzwerke (1)
- multi core data processing (1)
- multi-class classification (1)
- natural disasters (1)
- natural language generation (1)
- navigation (1)
- neighborhood (1)
- networks-on-chip (1)
- neue Online-Fehlererkennungsmethode (1)
- neural networks (1)
- neutral endopeptidase (1)
- nichtlineare ICA (1)
- nichtlineare PCA (NLPCA) (1)
- nichtlineare Projektionen (1)
- non-monotonic reasoning (1)
- nonlinear ICA (1)
- nonlinear PCA (NLPCA) (1)
- nonlinear projections (1)
- nonphotorealistic rendering (NPR) (1)
- o-ambiguity (1)
- objective difficulty (1)
- omega (1)
- on-chip (1)
- on-farm evaluation (1)
- one-sided communication (1)
- oneM2M (1)
- oneM2M Ontology (1)
- ontologies (1)
- open source (1)
- optimization (1)
- organisational evolution (1)
- outlier detection (1)
- output space compaction (1)
- overcomplete ICA (1)
- pMOS radiation dosimeter (1)
- paper prototyping (1)
- parallel processing (1)
- parallel programming (1)
- parallel rewriting (1)
- parallel solving (1)
- parallele Programmierung (1)
- paralleles Lösen (1)
- parity aggregate operator (1)
- pattern recognition (1)
- pdf forms (1)
- philosophy of mathematics (1)
- physical Computing (1)
- physical computing tools (1)
- placement (1)
- planning (1)
- plug-ins (1)
- portfolio-based solving (1)
- prediction (1)
- predictive models (1)
- preferences (1)
- premise acceptability (1)
- priorities (1)
- probabilistic deep learning (1)
- probabilistic deep metric learning (1)
- probabilistische tiefe neuronale Netze (1)
- probabilistisches tiefes metrisches Lernen (1)
- process (1)
- process improvement (1)
- process model (1)
- process model alignment (1)
- process modeling (1)
- process modelling (1)
- process synchronization (1)
- professors (1)
- program encodings (1)
- programmed grammars (1)
- projection (1)
- proof (1)
- proof assistant (1)
- proof complexity (1)
- proof environment (1)
- propagation probability (1)
- prototype (1)
- pruritus (1)
- pulse stretching inverters (1)
- quality of life (1)
- quantum (1)
- radiation hardness (1)
- radiation hardness design (1)
- railway network (1)
- random forest (1)
- real arguments (1)
- real-time (1)
- real-time application (1)
- real-time mapping (1)
- real-walking (1)
- reconfiguration (1)
- reference (1)
- referential effectiveness (1)
- regression (1)
- regular language (1)
- rekonfigurierbar (1)
- relevance (1)
- reliability analysis (1)
- reliability assessment (1)
- repair (1)
- resources (1)
- restricted parallelism (1)
- risk analysis (1)
- robust ICA (1)
- robuste ICA (1)
- safety (1)
- satisfiability (1)
- scheduling (1)
- search (1)
- secondary computer science education (1)
- segmentation (1)
- selbstanpassendes Multiprozessorsystem (1)
- selbstprüfende Schaltungen (1)
- selective fault tolerance (1)
- self-checking (1)
- semantic domain modeling (1)
- semantic ranking (1)
- semantic search (1)
- semantic search evaluation (1)
- semantic search methods (1)
- semantic web (1)
- semantische Domänenmodellierung (1)
- semantische Suche (1)
- semantisches Netz (1)
- semantisches Ranking (1)
- sensitivity (1)
- service composition (1)
- shader (1)
- simplicity (1)
- single event upsets (1)
- single-event transient (1)
- single-trial-analysis (1)
- site-specific weed management (1)
- skeletonization (1)
- sleep quality (1)
- smart farming (1)
- smartphone (1)
- socio-technical system (1)
- soft errors (1)
- software (1)
- software engineering (1)
- software-based cache coherence (1)
- sozio-technisches System (1)
- space missions (1)
- speed independence (1)
- stable model semantics (1)
- state complexity (1)
- static analysis (1)
- static prediction games (1)
- statistics program R (1)
- strahleninduzierte Einzelereignis-Effekte (1)
- strong equivalence (1)
- structured output prediction (1)
- strukturierte Vorhersage (1)
- study problems (1)
- stylization (1)
- sufficiency (1)
- suicidal ideations (1)
- supply chains (1)
- support vector machines (1)
- synthesis (1)
- systematic (1)
- systematisch (1)
- tableau calculi (1)
- tactic (1)
- teachers (1)
- teaching (1)
- technical notes and rapid communications (1)
- tele-teaching (1)
- temporary binding (1)
- terrain models (1)
- test response compaction (1)
- theory (1)
- theory of computation (1)
- think aloud (1)
- timing (1)
- tools (1)
- tools for teaching (1)
- topics (1)
- touch input (1)
- tptp (1)
- transformation (1)
- transient Faults (1)
- transient analysis (1)
- triangulated irregular networks (1)
- triple modular redundancy (1)
- tutorial section (1)
- unfounded sets (1)
- unidirektionale Fehler (1)
- university education (1)
- user experience (1)
- video annotation (1)
- virtual 3D city model (1)
- virtual machine (1)
- weather extremes (1)
- wheat crops (1)
- work productivity (1)
- workflow management (1)
- xAPI (1)
- yellow rust (1)
- zero-aliasing (1)
- überbestimmte ICA (1)
Institute
- Institut für Informatik und Computational Science (998)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (18)
- Extern (5)
- Institut für Physik und Astronomie (2)
- Universitätsbibliothek (2)
- Zentrum für Qualitätsentwicklung in Lehre und Studium (ZfQ) (2)
- eLiS - E-Learning in Studienbereichen (2)
- Department Erziehungswissenschaft (1)
- Department Linguistik (1)
- Historisches Institut (1)
Learning how to prove
(2018)
We have developed an alternative approach to teaching computer science students how to prove. First, students are taught how to prove theorems with the Coq proof assistant. In a second, more difficult, step students will transfer their acquired skills to the area of textbook proofs. In this article we present a realisation of the second step. Proofs in Coq have a high degree of formality while textbook proofs have only a medium one. Therefore our key idea is to reduce the degree of formality from the level of Coq to textbook proofs in several small steps. For that purpose we introduce three proof styles between Coq and textbook proofs, called line by line comments, weakened line by line comments, and structure faithful proofs. While this article is mostly conceptional we also report on experiences with putting our approach into practise.
This paper proposes an education approach for master and bachelor students to enhance their skills in the area of reliability, safety and security of the electronic components in automated driving. The approach is based on the active synergetic work of research institutes, academia and industry in the frame of joint lab. As an example, the jointly organized summer school with the respective focus is organized and elaborated.
We propose a new temporal extension of the logic of Here-and-There (HT) and its equilibria obtained by combining it with dynamic logic over (linear) traces. Unlike previous temporal extensions of HT based on linear temporal logic, the dynamic logic features allow us to reason about the composition of actions. For instance, this can be used to exercise fine grained control when planning in robotics, as exemplified by GOLOG. In this paper, we lay the foundations of our approach, and refer to it as Linear Dynamic Equilibrium Logic, or simply DEL. We start by developing the formal framework of DEL and provide relevant characteristic results. Among them, we elaborate upon the relationships to traditional linear dynamic logic and previous temporal extensions of HT.
We present a prototype of an integrated reasoning environment for educational purposes. The presented tool is a fragment of a proof assistant and automated theorem prover. We describe the existing and planned functionality of the theorem prover and especially the functionality of the educational fragment. This currently supports working with terms of the untyped lambda calculus and addresses both undergraduate students and researchers. We show how the tool can be used to support the students' understanding of functional programming and discuss general problems related to the process of building theorem proving software that aims at supporting both research and education.
Manufacturing industries are undergoing a major paradigm shift towards more autonomy. Automated planning and scheduling then becomes a necessity. The Planning and Execution Competition for Logistics Robots in Simulation held at ICAPS is based on this scenario and provides an interesting testbed. However, the posed problem is challenging as also demonstrated by the somewhat weak results in 2017. The domain requires temporal reasoning and dealing with uncertainty. We propose a novel planning system based on Answer Set Programming and the Clingo solver to tackle these problems and incentivize robot cooperation. Our results show a significant performance improvement, both, in terms of lowering computational requirements and better game metrics.
Scenograph
(2018)
When developing a real-walking virtual reality experience, designers generally create virtual locations to fit a specific tracking volume. Unfortunately, this prevents the resulting experience from running on a smaller or differently shaped tracking volume. To address this, we present a software system called Scenograph. The core of Scenograph is a tracking volume-independent representation of real-walking experiences. Scenograph instantiates the experience to a tracking volume of given size and shape by splitting the locations into smaller ones while maintaining narrative structure. In our user study, participants' ratings of realism decreased significantly when existing techniques were used to map a 25m2 experience to 9m2 and an L-shaped 8m2 tracking volume. In contrast, ratings did not differ when Scenograph was used to instantiate the experience.
Beware of SMOMBIES
(2018)
Several research evaluated the user's style of walking for the verification of a claimed identity and showed high authentication accuracies in many settings. In this paper we present a system that successfully verifies a user's identity based on many real world smartphone placements and yet not regarded interactions while walking. Our contribution is the distinction of all considered activities into three distinct subsets and a specific one-class Support Vector Machine per subset. Using sensor data of 30 participants collected in a semi-supervised study approach, we prove that unsupervised verification is possible with very low false-acceptance and false-rejection rates. We furthermore show that these subsets can be distinguished with a high accuracy and demonstrate that this system can be deployed on off-the-shelf smartphones.
Die stetige Weiterentwicklung von VR-Systemen bietet neue Möglichkeiten der Interaktion mit virtuellen Objekten im dreidimensionalen Raum, stellt Entwickelnde von VRAnwendungen aber auch vor neue Herausforderungen. Selektions- und Manipulationstechniken müssen unter Berücksichtigung des Anwendungsszenarios, der Zielgruppe und der zur Verfügung stehenden Ein- und Ausgabegeräte ausgewählt werden. Diese Arbeit leistet einen Beitrag dazu, die Auswahl von passenden Interaktionstechniken zu unterstützen. Hierfür wurde eine repräsentative Menge von Selektions- und Manipulationstechniken untersucht und, unter Berücksichtigung existierender Klassifikationssysteme, eine Taxonomie entwickelt, die die Analyse der Techniken hinsichtlich interaktionsrelevanter Eigenschaften ermöglicht. Auf Basis dieser Taxonomie wurden Techniken ausgewählt, die in einer explorativen Studie verglichen wurden, um Rückschlüsse auf die Dimensionen der Taxonomie zu ziehen und neue Indizien für Vor- und Nachteile der Techniken in spezifischen Anwendungsszenarien zu generieren. Die Ergebnisse der Arbeit münden in eine Webanwendung, die Entwickelnde von VR-Anwendungen gezielt dabei unterstützt, passende Selektions- und Manipulationstechniken für ein Anwendungsszenario auszuwählen, indem Techniken auf Basis der Taxonomie gefiltert und unter Verwendung der Resultate aus der Studie sortiert werden können.
In computer science, computer systems are both, objects of investigation and tools that enable creative learning and design. Tools for learning have a long tradition in computer science education. Already in the late 1960s, Papert developed a concept which had an immense impact on the development of informal education in the following years: his theory of constructionism understands learning as a creative process of knowledge construction that is most effective when learners create something purposeful that they can try out, show around, discuss, analyse and receive praise for. By now, there are numerous learning and programming environments that are based on the constructionist ideas. Modern tools offer opportunities for students to learn in motivating ways and gain impressive results in programming games, animations, implementing 3D models or developing interactive objects. This article gives an overview of computer science education research related to tools and media to be used in educational settings. We analyse different types of tools with a special focus on the categorization and development of tools for student adequate physical computing activities in the classroom. Research around the development and evaluation of tools and learning resources in the domain of physical computing is illustrated with the example of "My Interactive Garden", a constructionist learning and programming environment. It is explained how the results from empirical studies are integrated in the continuous development of the learning material.
The Internet of Things (IoT) is a system of physical objects that can be discovered, monitored, controlled, or interacted with by electronic devices that communicate over various networking interfaces and eventually can be connected to the wider Internet. [Guinard and Trifa, 2016]. IoT devices are equipped with sensors and/or actuators and may be constrained in terms of memory, computational power, network bandwidth, and energy. Interoperability can help to manage such heterogeneous devices. Interoperability is the ability of different types of systems to work together smoothly. There are four levels of interoperability: physical, network and transport, integration, and data. The data interoperability is subdivided into syntactic and semantic data. Semantic data describes the meaning of data and the common understanding of vocabulary e.g. with the help of dictionaries, taxonomies, ontologies. To achieve interoperability, semantic interoperability is necessary.
Many organizations and companies are working on standards and solutions for interoperability in the IoT. However, the commercial solutions produce a vendor lock-in. They focus on centralized approaches such as cloud-based solutions. This thesis proposes a decentralized approach namely Edge Computing. Edge Computing is based on the concepts of mesh networking and distributed processing. This approach has an advantage that information collection and processing are placed closer to the sources of this information. The goals are to reduce traffic, latency, and to be robust against a lossy or failed Internet connection.
We see management of IoT devices from the network configuration management perspective. This thesis proposes a framework for network configuration management of heterogeneous, constrained IoT devices by using semantic descriptions for interoperability. The MYNO framework is an acronym for MQTT, YANG, NETCONF and Ontology. The NETCONF protocol is the IETF standard for network configuration management. The MQTT protocol is the de-facto standard in the IoT. We picked up the idea of the NETCONF-MQTT bridge, originally proposed by Scheffler and Bonneß[2017], and extended it with semantic device descriptions. These device descriptions provide a description of the device capabilities. They are based on the oneM2M Base ontology and formalized by the Semantic Web Standards.
The novel approach is using a ontology-based device description directly on a constrained device in combination with the MQTT protocol. The bridge was extended in order to query such descriptions. Using a semantic annotation, we achieved that the device capabilities are self-descriptive, machine readable and re-usable.
The concept of a Virtual Device was introduced and implemented, based on semantic device descriptions. A Virtual Device aggregates the capabilities of all devices at the edge network and contributes therefore to the scalability. Thus, it is possible to control all devices via a single RPC call.
The model-driven NETCONF Web-Client is generated automatically from this YANG model which is generated by the bridge based on the semantic device description. The Web-Client provides a user-friendly interface, offers RPC calls and displays sensor values. We demonstrate the feasibility of this approach in different use cases: sensor and actuator scenarios, as well as event configuration and triggering.
The semantic approach results in increased memory overhead. Therefore, we evaluated CBOR and RDF HDT for optimization of ontology-based device descriptions for use on constrained devices. The evaluation shows that CBOR is not suitable for long strings and RDF HDT is a promising candidate but is still a W3C Member Submission. Finally, we used an optimized JSON-LD format for the syntax of the device descriptions.
One of the security tasks of network management is the distribution of firmware updates. The MYNO Update Protocol (MUP) was developed and evaluated on constrained devices CC2538dk and 6LoWPAN. The MYNO update process is focused on freshness and authenticity of the firmware. The evaluation shows that it is challenging but feasible to bring the firmware updates to constrained devices using MQTT. As a new requirement for the next MQTT version, we propose to add a slicing feature for the better support of constrained devices. The MQTT broker should slice data to the maximum packet size specified by the device and transfer it slice-by-slice.
For the performance and scalability evaluation of MYNO framework, we setup the High Precision Agriculture demonstrator with 10 ESP-32 NodeMCU boards at the edge of the network. The ESP-32 NodeMCU boards, connected by WLAN, were equipped with six sensors and two actuators. The performance evaluation shows that the processing of ontology-based descriptions on a Raspberry Pi 3B with the RDFLib is a challenging task regarding computational power. Nevertheless, it is feasible because it must be done only once per device during the discovery process.
The MYNO framework was tested with heterogeneous devices such as CC2538dk from Texas Instruments, Arduino Yún Rev 3, and ESP-32 NodeMCU, and IP-based networks such as 6LoWPAN and WLAN.
Summarizing, with the MYNO framework we could show that the semantic approach on constrained devices is feasible in the IoT.
Nowadays, business processes are increasingly supported by IT services that produce massive amounts of event data during the execution of a process. These event data can be used to analyze the process using process mining techniques to discover the real process, measure conformance to a given process model, or to enhance existing models with performance information. Mapping the produced events to activities of a given process model is essential for conformance checking, annotation and understanding of process mining results. In order to accomplish this mapping with low manual effort, we developed a semi-automatic approach that maps events to activities using insights from behavioral analysis and label analysis. The approach extracts Declare constraints from both the log and the model to build matching constraints to efficiently reduce the number of possible mappings. These mappings are further reduced using techniques from natural language processing, which allow for a matching based on labels and external knowledge sources. The evaluation with synthetic and real-life data demonstrates the effectiveness of the approach and its robustness toward non-conforming execution logs.
Novel two-dimensional tactile displays enable blind users to not only get access to the textual but also to the graphical content of a graphical user interface. Due to the higher amount of information that can be presented in parallel, orientation and exploration can be more complex. In this paper we present the HyperBraille system, which consists of a pin-matrix device as well as a graphical screen reader providing the user with appropriate presentation and interaction possibilities. To allow for a detailed analysis of bimanual interaction strategies on a pin-matrix device, we conducted two user studies with a total of 12 blind people. The task was to fill in .pdf forms on the pin-matrix device by using different input methods, namely gestures, built-in hardware buttons as well as a conventional PC keyboard. The forms were presented in a semigraphic view type that not only contains Braille but also tactile widgets in a spatial arrangement. While completion time and error rate partly depended on the chosen input method, the usage of special reading strategies seemed to be independent of it. A direct comparison of the system and a conventional assistive technology (screen reader with single-line Braille device) showed that interaction on the pin-matrix device can be very efficient if the user is trained. The two-dimensional output can improve access to .pdf forms with insufficient accessibility as the mapping of input controls and the corresponding labels can be supported by a spatial presentation.
Automated storage and retrieval systems are principal components of modern production and warehouse facilities. In particular, automated guided vehicles nowadays substitute human-operated pallet trucks in transporting production materials between storage locations and assembly stations. While low-level control systems take care of navigating such driverless vehicles along programmed routes and avoid collisions even under unforeseen circumstances, in the common case of multiple vehicles sharing the same operation area, the problem remains how to set up routes such that a collection of transport tasks is accomplished most effectively. We address this prevalent problem in the context of car assembly at Mercedes-Benz Ludwigsfelde GmbH, a large-scale producer of commercial vehicles, where routes for automated guided vehicles used in the production process have traditionally been hand-coded by human engineers. Such adhoc methods may suffice as long as a running production process remains in place, while any change in the factory layout or production targets necessitates tedious manual reconfiguration, not to mention the missing portability between different production plants. Unlike this, we propose a declarative approach based on Answer Set Programming to optimize the routes taken by automated guided vehicles for accomplishing transport tasks. The advantages include a transparent and executable problem formalization, provable optimality of routes relative to objective criteria, as well as elaboration tolerance towards particular factory layouts and production targets. Moreover, we demonstrate that our approach is efficient enough to deal with the transport tasks evolving in realistic production processes at the car factory of Mercedes-Benz Ludwigsfelde GmbH.
Research publications and data nowadays should be publicly available on the internet and, theoretically, usable for everyone to develop further research, products, or services. The long-term accessibility of research data is, therefore, fundamental in the economy of the research production process. However, the availability of data is not sufficient by itself, but also their quality must be verifiable. Measures to ensure reuse and reproducibility need to include the entire research life cycle, from the experimental design to the generation of data, quality control, statistical analysis, interpretation, and validation of the results. Hence, high-quality records, particularly for providing a string of documents for the verifiable origin of data, are essential elements that can act as a certificate for potential users (customers). These records also improve the traceability and transparency of data and processes, therefore, improving the reliability of results. Standards for data acquisition, analysis, and documentation have been fostered in the last decade driven by grassroot initiatives of researchers and organizations such as the Research Data Alliance (RDA). Nevertheless, what is still largely missing in the life science academic research are agreed procedures for complex routine research workflows. Here, well-crafted documentation like standard operating procedures (SOPs) offer clear direction and instructions specifically designed to avoid deviations as an absolute necessity for reproducibility. Therefore, this paper provides a standardized workflow that explains step by step how to write an SOP to be used as a starting point for appropriate research documentation.
Research publications and data nowadays should be publicly available on the internet and, theoretically, usable for everyone to develop further research, products, or services. The long-term accessibility of research data is, therefore, fundamental in the economy of the research production process. However, the availability of data is not sufficient by itself, but also their quality must be verifiable. Measures to ensure reuse and reproducibility need to include the entire research life cycle, from the experimental design to the generation of data, quality control, statistical analysis, interpretation, and validation of the results. Hence, high-quality records, particularly for providing a string of documents for the verifiable origin of data, are essential elements that can act as a certificate for potential users (customers). These records also improve the traceability and transparency of data and processes, therefore, improving the reliability of results. Standards for data acquisition, analysis, and documentation have been fostered in the last decade driven by grassroot initiatives of researchers and organizations such as the Research Data Alliance (RDA). Nevertheless, what is still largely missing in the life science academic research are agreed procedures for complex routine research workflows. Here, well-crafted documentation like standard operating procedures (SOPs) offer clear direction and instructions specifically designed to avoid deviations as an absolute necessity for reproducibility. Therefore, this paper provides a standardized workflow that explains step by step how to write an SOP to be used as a starting point for appropriate research documentation.
Institutionelle Bildung ist für autistische Lernende mit vielgestaltigen und spezifischen Hindernissen verbunden. Dies gilt insbesondere im Zusammenhang mit Inklusion, deren Relevanz nicht zuletzt durch das Übereinkommen der Vereinten Nationen über die Rechte von Menschen mit Behinderung gegeben ist.
Diese Arbeit diskutiert zahlreiche lernrelevante Besonderheiten im Kontext von Autismus und zeigt Diskrepanzen zu den nicht immer ausreichend angemessenen institutionellen Lehrkonzepten. Eine zentrale These ist hierbei, dass die ungewöhnlich intensive Aufmerksamkeit von Autist*innen für ihre Spezialinteressen dafür genutzt werden kann, das Lernen mit fremdgestellten Inhalten zu erleichtern. Darauf aufbauend werden Lösungsansätze diskutiert, welche in einem neuartigen Konzept für ein digitales mehrgerätebasiertes Lernspiel resultieren.
Eine wesentliche Herausforderung bei der Konzeption spielbasierten Lernens besteht in der adäquaten Einbindung von Lerninhalten in einen fesselnden narrativen Kontext. Am Beispiel von Übungen zur emotionalen Deutung von Mimik, welche für das Lernen von sozioemotionalen Kompetenzen besonders im Rahmen von Therapiekonzepten bei Autismus Verwendung finden, wird eine angemessene Narration vorgestellt, welche die störungsarme Einbindung dieser sehr speziellen Lerninhalte ermöglicht.
Die Effekte der einzelnen Konzeptionselemente werden anhand eines prototypisch entwickelten Lernspiels untersucht. Darauf aufbauend zeigt eine quantitative Studie die gute Akzeptanz und Nutzerfreundlichkeit des Spiels und belegte vor allem die
Verständlichkeit der Narration und der Spielelemente. Ein weiterer Schwerpunkt liegt in der minimalinvasiven Untersuchung möglicher Störungen des Spielerlebnisses durch den Wechsel zwischen verschiedenen Endgeräten, für die ein innovatives Messverfahren entwickelt wurde.
Im Ergebnis beleuchtet diese Arbeit die Bedeutung und die Grenzen von spielbasierten Ansätzen für autistische Lernende. Ein großer Teil der vorgestellten Konzepte lässt sich auf andersartige Lernszenarien übertragen. Das dafür entwickelte technische Framework zur Realisierung narrativer Lernpfade ist ebenfalls darauf vorbereitet, für weitere Lernszenarien, gerade auch im institutionellen Kontext, Verwendung zu finden.
The soft error rate (SER) due to heavy-ion irradiation of a clock tree is investigated in this paper. A method for clock tree SER prediction is developed, which employs a dedicated soft error analysis tool to characterize the single-event transient (SET) sensitivities of clock inverters and other commercial tools to calculate the SER through fault-injection simulations. A test circuit including a flip-flop chain and clock tree in a 65 nm CMOS technology is developed through the automatic ASIC design flow. This circuit is analyzed with the developed method to calculate its clock tree SER. In addition, this circuit is implemented in a 65 nm test chip and irradiated by heavy ions to measure its SER resulting from the SETs in the clock tree. The experimental and calculation results of this case study present good correlation, which verifies the effectiveness of the developed method.
The aim of our project design space exploration with answer set programming is to develop a general framework based on Answer Set Programming (ASP) that finds valid solutions to the system design problem and simultaneously performs Design Space Exploration (DSE) to find the most favorable alternatives. We leverage recent developments in ASP solving that allow for tight integration of background theories to create a holistic framework for effective DSE.
The Potsdam answer set solving collection, or Potassco for short, bundles various tools implementing and/or applying answer set programming. The article at hand succeeds an earlier description of the Potassco project published in Gebser et al. (AI Commun 24(2):107-124, 2011). Hence, we concentrate in what follows on the major features of the most recent, fifth generation of the ASP system clingo and highlight some recent resulting application systems.
Answer Set Programming faces an increasing popularity for problem solving in various domains. While its modeling language allows us to express many complex problems in an easy way, its solving technology enables their effective resolution. In what follows, we detail some of the key factors of its success. Answer Set Programming [ASP; Brewka et al. Commun ACM 54(12):92–103, (2011)] is seeing a rapid proliferation in academia and industry due to its easy and flexible way to model and solve knowledge-intense combinatorial (optimization) problems. To this end, ASP offers a high-level modeling language paired with high-performance solving technology. As a result, ASP systems provide out-off-the-box, general-purpose search engines that allow for enumerating (optimal) solutions. They are represented as answer sets, each being a set of atoms representing a solution. The declarative approach of ASP allows a user to concentrate on a problem’s specification rather than the computational means to solve it. This makes ASP a prime candidate for rapid prototyping and an attractive tool for teaching key AI techniques since complex problems can be expressed in a succinct and elaboration tolerant way. This is eased by the tuning of ASP’s modeling language to knowledge representation and reasoning (KRR). The resulting impact is nicely reflected by a growing range of successful applications of ASP [Erdem et al. AI Mag 37(3):53–68, 2016; Falkner et al. Industrial applications of answer set programming. K++nstliche Intelligenz (2018)]
THIS INSTALLMENT OF Research for Practice provides curated reading guides to technology for underserved communities and to new developments in personal fabrication. First, Tawanna Dillahunt describes design considerations and technology for underserved and impoverished communities. Designing for the more than 1.6 billion impoverished individuals worldwide requires special consideration of community needs, constraints, and context. Her selections span protocols for poor-quality communication networks, community-driven content generation, and resource and public service discovery. Second, Stefanie Mueller and Patrick Baudisch provide an overview of recent advances in personal fabrication (for example, 3D printers).
In vielen Studiengängen kommt es durch die oft heterogenen Vorkenntnisse in der Studieneingangsphase zu mangelnder Motivation durch Über- oder Unterforderung. Dieses Problem tritt auch in der musiktheoretischen Grundausbildung an Hochschulen auf. Durch Einsatz von Elementen, die aus dem Unterhaltungskontext geläufig sind, kann eine Steigerung der Motivation erreicht werden. Die Nutzung solcher Elemente wird als Gamification bezeichnet.
Das Ziel der vorliegenden Arbeit ist es, am Fallbeispiel der musiktheoretischen Grundausbildung zu analysieren, ob Lerngelegenheiten durch einen gamifizierten interaktiven Prototyp einer Lernumgebung unterstützt werden können. Dazu wird die folgende Forschungsfrage gestellt: Inwieweit wirkt Gamification auf die Motivation bei den Lernenden zur Beschäftigung mit dem Thema (musikalische) Funktionsanalyse?
Um die Forschungsfragen zu beantworten, wurde zunächst ein systematisches, theoriegeleitetes Vorgehensmodell zur Gamification von Lernumgebungen entwickelt und angewandt. Der so entstandene Prototyp wurde anschließend um alle Game-Design-Elemente reduziert und im Rahmen einer experimentellen Studie mit zwei unabhängigen Versuchsgruppen mit der gamifizierten Variante verglichen.
Die Untersuchung zeigte, dass die Gamification einer Lernanwendung nach dem entwickelten Vorgehensmodell grundsätzlich das Potenzial besitzt, manche Aspekte des Nutzungserlebnisses (UX) positiv zu beeinflussen. Insbesondere hatte die Gamification positive Effekte auf die Joy of Use und die Immersivität. Allerdings blieb das Ausmaß der beobachteten Effekte deutlich hinter den Erwartungen zurück, die auf Basis verschiedener Motivationstheorien getroffen wurden.
Daher erscheint Gamification besonders in außeruniversitären Kontexten vielversprechend, in denen der Fokus auf einer Erhöhung der Joy of Use oder einer Steigerung der Immersivität liegt. Allerdings lassen sich durch die Untersuchung neue Erkenntnisse zur emotionalen Wirkung von Gamification und zu einem systematischen Vorgehen bei der Gamification von Lernanwendungen herausstellen.
Weiterführende Forschung könnte an diese Erkenntnisse anknüpfen, indem sie die emotionale Wirkung von Gamification und deren Einfluss auf die Motivation näher untersucht. Darüber hinaus sollte sie Gamification auch aus einer entscheidungstheoretischen Perspektive betrachten und Analysemethoden entwickeln, mit denen entschieden werden kann, ob der Einsatz von Gamification zur Motivationssteigerung in einem spezifischen Anwendungsfall zielführend ist. Unter Verwendung des entwickelten Vorgehensmodells kann es sinnvoll sein, näher zu untersuchen, welche Faktoren insgesamt für das Gelingen einer Gamification-Maßnahme in Bildungskontexten entscheidend sind. Die Erkenntnisse einer solchen Untersuchung könnten entscheidend zur Verbesserung und Validierung des Vorgehensmodells beitragen.
In the last decades, there was a notable progress in solving the well-known Boolean satisfiability (Sat) problem, which can be witnessed by powerful Sat solvers. One of the reasons why these solvers are so fast are structural properties of instances that are utilized by the solver’s interna. This thesis deals with the well-studied structural property treewidth, which measures the closeness of an instance to being a tree. In fact, there are many problems parameterized by treewidth that are solvable in polynomial time in the instance size when parameterized by treewidth.
In this work, we study advanced treewidth-based methods and tools for problems in knowledge representation and reasoning (KR). Thereby, we provide means to establish precise runtime results (upper bounds) for canonical problems relevant to KR. Then, we present a new type of problem reduction, which we call decomposition-guided (DG) that
allows us to precisely monitor the treewidth when reducing from one problem to another problem. This new reduction type will be the basis for a long-open lower bound result for quantified Boolean formulas and allows us to design a new methodology for establishing runtime lower bounds for problems parameterized by treewidth.
Finally, despite these lower bounds, we provide an efficient implementation of algorithms that adhere to treewidth. Our approach finds suitable abstractions of instances, which are subsequently refined in a recursive fashion, and it uses Sat solvers for solving subproblems. It turns out that our resulting solver is quite competitive for two canonical counting problems related to Sat.
Discriminative Models for Biometric Identification using Micro- and Macro-Movements of the Eyes
(2021)
Human visual perception is an active process. Eye movements either alternate between fixations and saccades or follow a smooth pursuit movement in case of moving targets. Besides these macroscopic gaze patterns, the eyes perform involuntary micro-movements during fixations which are commonly categorized into micro-saccades, drift and tremor. Eye movements are frequently studied in cognitive psychology, because they reflect a complex interplay of perception, attention and oculomotor control.
A common insight of psychological research is that macro-movements are highly individual. Inspired by this finding, there has been a considerable amount of prior research on oculomotoric biometric identification. However, the accuracy of known approaches is too low and the time needed for identification is too long for any practical application. This thesis explores discriminative models for the task of biometric identification.
Discriminative models optimize a quality measure of the predictions and are usually superior to generative approaches in discriminative tasks. However, using discriminative models requires to select a suitable form of data representation for sequential eye gaze data; i.e., by engineering features or constructing a sequence kernel and the performance of the classification model strongly depends on the data representation. We study two fundamentally different ways of representing eye gaze within a discriminative framework. In the first part of this thesis, we explore the integration of data and psychological background knowledge in the form of generative models to construct representations. To this end, we first develop generative statistical models of gaze behavior during reading and scene viewing that account for viewer-specific distributional properties of gaze patterns. In a second step, we develop a discriminative identification model by deriving Fisher kernel functions from these and several baseline models. We find that an SVM with Fisher kernel is able to reliably identify users based on their eye gaze during reading and scene viewing. However, since the generative models are constrained to use low-frequency macro-movements, they discard a significant amount of information contained in the raw eye tracking signal at a high cost: identification requires about one minute of input recording, which makes it inapplicable for real world biometric systems. In the second part of this thesis, we study a purely data-driven modeling approach. Here, we aim at automatically discovering the individual pattern hidden in the raw eye tracking signal. To this end, we develop a deep convolutional neural network DeepEyedentification that processes yaw and pitch gaze velocities and learns a representation end-to-end. Compared to prior work, this model increases the identification accuracy by one order of magnitude and the time to identification decreases to only seconds. The DeepEyedentificationLive model further improves upon the identification performance by processing binocular input and it also detects presentation-attacks.
We find that by learning a representation, the performance of oculomotoric identification and presentation-attack detection can be driven close to practical relevance for biometric applications. Eye tracking devices with high sampling frequency and precision are expensive and the applicability of eye movement as a biometric feature heavily depends on cost of recording devices.
In the last part of this thesis, we therefore study the requirements on data quality by evaluating the performance of the DeepEyedentificationLive network under reduced spatial and temporal resolution. We find that the method still attains a high identification accuracy at a temporal resolution of only 250 Hz and a precision of 0.03 degrees. Reducing both does not have an additive deteriorating effect.
In this paper, we consider the computational power of a new variant of networks of splicing processors in which each processor as well as the data navigating throughout the network are now considered to be polarized. While the polarization of every processor is predefined (negative, neutral, positive), the polarization of data is dynamically computed by means of a valuation mapping. Consequently, the protocol of communication is naturally defined by means of this polarization. We show that networks of polarized splicing processors (NPSP) of size 2 are computationally complete, which immediately settles the question of designing computationally complete NPSPs of minimal size. With two more nodes we can simulate every nondeterministic Turing machine without increasing the time complexity. Particularly, we prove that NPSP of size 4 can accept all languages in NP in polynomial time. Furthermore, another computational model that is universal, namely the 2-tag system, can be simulated by NPSP of size 3 preserving the time complexity. All these results can be obtained with NPSPs with valuations in the set as well. We finally show that Turing machines can simulate a variant of NPSPs and discuss the time complexity of this simulation.
The business problem of having inefficient processes, imprecise process analyses, and simulations as well as non-transparent artificial neuronal network models can be overcome by an easy-to-use modeling concept. With the aim of developing a flexible and efficient approach to modeling, simulating, and optimizing processes, this paper proposes a flexible Concept of Neuronal Modeling (CoNM). The modeling concept, which is described by the modeling language designed and its mathematical formulation and is connected to a technical substantiation, is based on a collection of novel sub-artifacts. As these have been implemented as a computational model, the set of CoNM tools carries out novel kinds of Neuronal Process Modeling (NPM), Neuronal Process Simulations (NPS), and Neuronal Process Optimizations (NPO). The efficacy of the designed artifacts was demonstrated rigorously by means of six experiments and a simulator of real industrial production processes.
Metabolic networks play a crucial role in biology since they capture all chemical reactions in an organism. While there are networks of high quality for many model organisms, networks for less studied organisms are often of poor quality and suffer from incompleteness. To this end, we introduced in previous work an answer set programming (ASP)-based approach to metabolic network completion. Although this qualitative approach allows for restoring moderately degraded networks, it fails to restore highly degraded ones. This is because it ignores quantitative constraints capturing reaction rates. To address this problem, we propose a hybrid approach to metabolic network completion that integrates our qualitative ASP approach with quantitative means for capturing reaction rates. We begin by formally reconciling existing stoichiometric and topological approaches to network completion in a unified formalism. With it, we develop a hybrid ASP encoding and rely upon the theory reasoning capacities of the ASP system dingo for solving the resulting logic program with linear constraints over reals. We empirically evaluate our approach by means of the metabolic network of Escherichia coli. Our analysis shows that our novel approach yields greatly superior results than obtainable from purely qualitative or quantitative approaches.
In recent years, named entity linking (NEL) tools were primarily developed in terms of a general approach, whereas today numerous tools are focusing on specific domains such as e.g. the mapping of persons and organizations only, or the annotation of locations or events in microposts. However, the available benchmark datasets necessary for the evaluation of NEL tools do not reflect this focalizing trend. We have analyzed the evaluation process applied in the NEL benchmarking framework GERBIL [in: Proceedings of the 24th International Conference on World Wide Web (WWW’15), International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, Switzerland, 2015, pp. 1133–1143, Semantic Web 9(5) (2018), 605–625] and all its benchmark datasets. Based on these insights we have extended the GERBIL framework to enable a more fine grained evaluation and in depth analysis of the available benchmark datasets with respect to different emphases. This paper presents the implementation of an adaptive filter for arbitrary entities and customized benchmark creation as well as the automated determination of typical NEL benchmark dataset properties, such as the extent of content-related ambiguity and diversity. These properties are integrated on different levels, which also enables to tailor customized new datasets out of the existing ones by remixing documents based on desired emphases. Besides a new system library to enrich provided NIF [in: International Semantic Web Conference (ISWC’13), Lecture Notes in Computer Science, Vol. 8219, Springer, Berlin, Heidelberg, 2013, pp. 98–113] datasets with statistical information, best practices for dataset remixing are presented, and an in depth analysis of the performance of entity linking systems on special focus datasets is presented.
Detect me if you can
(2019)
Spam Bots have become a threat to online social networks with their malicious behavior, posting misinformation messages and influencing online platforms to fulfill their motives. As spam bots have become more advanced over time, creating algorithms to identify bots remains an open challenge. Learning low-dimensional embeddings for nodes in graph structured data has proven to be useful in various domains. In this paper, we propose a model based on graph convolutional neural networks (GCNN) for spam bot detection. Our hypothesis is that to better detect spam bots, in addition to defining a features set, the social graph must also be taken into consideration. GCNNs are able to leverage both the features of a node and aggregate the features of a node’s neighborhood. We compare our approach, with two methods that work solely on a features set and on the structure of the graph. To our knowledge, this work is the first attempt of using graph convolutional neural networks in spam bot detection.
In this paper, we consider counting and projected model counting of extensions in abstract argumentation for various semantics. When asking for projected counts we are interested in counting the number of extensions of a given argumentation framework while multiple extensions that are identical when restricted to the projected arguments count as only one projected extension. We establish classical complexity results and parameterized complexity results when the problems are parameterized by treewidth of the undirected argumentation graph. To obtain upper bounds for counting projected extensions, we introduce novel algorithms that exploit small treewidth of the undirected argumentation graph of the input instance by dynamic programming (DP). Our algorithms run in time double or triple exponential in the treewidth depending on the considered semantics. Finally, we take the exponential time hypothesis (ETH) into account and establish lower bounds of bounded treewidth algorithms for counting extensions and projected extension.
A distinguishing feature of Answer Set Programming is that all atoms belonging to a stable model must be founded. That is, an atom must not only be true but provably true. This can be made precise by means of the constructive logic of Here-and-There, whose equilibrium models correspond to stable models. One way of looking at foundedness is to regard Boolean truth values as ordered by letting true be greater than false. Then, each Boolean variable takes the smallest truth value that can be proven for it. This idea was generalized by Aziz to ordered domains and applied to constraint satisfaction problems. As before, the idea is that a, say integer, variable gets only assigned to the smallest integer that can be justified. In this paper, we present a logical reconstruction of Aziz’ idea in the setting of the logic of Here-and-There. More precisely, we start by defining the logic of Here-and-There with lower bound founded variables along with its equilibrium models and elaborate upon its formal properties. Finally, we compare our approach with related ones and sketch future work.
Digitalisierung ermöglicht es uns, mit Partnern (z.B. Unternehmen, Institutionen) in einer IT-unterstützten Umgebung zu interagieren und Tätigkeiten auszuführen, die vormals manuell erledigt wurden. Ein Ziel der Digitalisierung ist dabei, Dienstleistungen unterschiedlicher fachlicher Domänen zu Prozessen zu kombinieren und vielen Nutzergruppen bedarfsgerecht zugänglich zu machen. Hierzu stellen Anbieter technische Dienste bereit, die in unterschiedliche Anwendungen integriert werden können.
Die Digitalisierung stellt die Anwendungsentwicklung vor neue Herausforderungen. Ein Aspekt ist die bedarfsgerechte Anbindung von Nutzern an Dienste. Zur Interaktion menschlicher Nutzer mit den Diensten werden Benutzungsschnittstellen benötigt, die auf deren Bedürfnisse zugeschnitten sind. Hierzu werden Varianten für spezifische Nutzergruppen (fachliche Varianten) und variierende Umgebungen (technische Varianten) benötigt. Zunehmend müssen diese mit Diensten anderer Anbieter kombiniert werden können, um domänenübergreifend Prozesse zu Anwendungen mit einem erhöhten Mehrwert für den Endnutzer zu verknüpfen (z.B. eine Flugbuchung mit einer optionalen Reiseversicherung).
Die Vielfältigkeit der Varianten lässt die Erstellung von Benutzungsschnittstellen komplex und die Ergebnisse sehr individuell erscheinen. Daher werden die Varianten in der Praxis vorwiegend manuell erstellt. Dies führt zur parallelen Entwicklung einer Vielzahl sehr ähnlicher Anwendungen, die nur geringes Potential zur Wiederverwendung besitzen. Die Folge sind hohe Aufwände bei Erstellung und Wartung. Dadurch wird häufig auf die Unterstützung kleiner Nutzerkreise mit speziellen Anforderungen verzichtet (z.B. Menschen mit physischen Einschränkungen), sodass diese weiterhin von der Digitalisierung ausgeschlossen bleiben.
Die Arbeit stellt eine konsistente Lösung für diese neuen Herausforderungen mit den Mitteln der modellgetriebenen Entwicklung vor. Sie präsentiert einen Ansatz zur Modellierung von Benutzungsschnittstellen, Varianten und Kompositionen und deren automatischer Generierung für digitale Dienste in einem verteilten Umfeld. Die Arbeit schafft eine Lösung zur Wiederverwendung und gemeinschaftlichen Nutzung von Benutzungsschnittstellen über Anbietergrenzen hinweg. Sie führt zu einer Infrastruktur, in der eine Vielzahl von Anbietern ihre Expertise in gemeinschaftliche Anwendungen einbringen können.
Die Beiträge bestehen im Einzelnen in Konzepten und Metamodellen zur Modellierung von Benutzungsschnittstellen, Varianten und Kompositionen sowie einem Verfahren zu deren vollständig automatisierten Transformation in funktionale Benutzungsschnittstellen. Zur Umsetzung der gemeinschaftlichen Nutzbarkeit werden diese ergänzt um eine universelle Repräsentation der Modelle, einer Methodik zur Anbindung unterschiedlicher Dienst-Anbieter sowie einer Architektur zur verteilten Nutzung der Artefakte und Verfahren in einer dienstorientierten Umgebung.
Der Ansatz bietet die Chance, unterschiedlichste Menschen bedarfsgerecht an der Digitalisierung teilhaben zu lassen. Damit setzt die Arbeit Impulse für zukünftige Methoden zur Anwendungserstellung in einem zunehmend vielfältigen Umfeld.
Zum Einfluss von Adaptivität auf die Wahrnehmung von Komplexität in der Mensch-Technik-Interaktion
(2021)
Wir leben in einer Gesellschaft, die von einem stetigen Wunsch nach Innovation und Fortschritt geprägt ist. Folgen dieses Wunsches sind die immer weiter fortschreitende Digitalisierung und informatische Vernetzung aller Lebensbereiche, die so zu immer komplexeren sozio-technischen Systemen führen. Ziele dieser Systeme sind u. a. die Unterstützung von Menschen, die Verbesserung ihrer Lebenssituation oder Lebensqualität oder die Erweiterung menschlicher Möglichkeiten. Doch haben neue komplexe technische Systeme nicht nur positive soziale und gesellschaftliche Effekte. Oft gibt es unerwünschte Nebeneffekte, die erst im Gebrauch sichtbar werden, und sowohl Konstrukteur*innen als auch Nutzer*innen komplexer vernetzter Technologien fühlen sich oft orientierungslos. Die Folgen können von sinkender Akzeptanz bis hin zum kompletten Verlust des Vertrauens in vernetze Softwaresysteme reichen. Da komplexe Anwendungen, und damit auch immer komplexere Mensch-Technik-Interaktionen, immer mehr an Relevanz gewinnen, ist es umso wichtiger, wieder Orientierung zu finden. Dazu müssen wir zuerst diejenigen Elemente identifizieren, die in der Interaktion mit vernetzten sozio-technischen Systemen zu Komplexität beitragen und somit Orientierungsbedarf hervorrufen.
Mit dieser Arbeit soll ein Beitrag geleistet werden, um ein strukturiertes Reflektieren über die Komplexität vernetzter sozio-technischer Systeme im gesamten Konstruktionsprozess zu ermöglichen. Dazu wird zuerst eine Definition von Komplexität und komplexen Systemen erarbeitet, die über das informatische Verständnis von Komplexität (also der Kompliziertheit von Problemen, Algorithmen oder Daten) hinausgeht. Im Vordergrund soll vielmehr die sozio-technische Interaktion mit und in komplexen vernetzten Systemen stehen. Basierend auf dieser Definition wird dann ein Analysewerkzeug entwickelt, welches es ermöglicht, die Komplexität in der Interaktion mit sozio-technischen Systemen sichtbar und beschreibbar zu machen.
Ein Bereich, in dem vernetzte sozio-technische Systeme zunehmenden Einzug finden, ist jener digitaler Bildungstechnologien. Besonders adaptiven Bildungstechnologien wurde in den letzten Jahrzehnten ein großes Potential zugeschrieben. Zwei adaptive Lehr- bzw. Trainingssysteme sollen deshalb exemplarisch mit dem in dieser Arbeit entwickelten Analysewerkzeug untersucht werden. Hierbei wird ein besonderes Augenmerkt auf den Einfluss von Adaptivität auf die Komplexität von Mensch-Technik-Interaktionssituationen gelegt. In empirischen Untersuchungen werden die Erfahrungen von Konstrukteur*innen und Nutzer*innen jener adaptiver Systeme untersucht, um so die entscheidenden Kriterien für Komplexität ermitteln zu können. Auf diese Weise können zum einen wiederkehrende Orientierungsfragen bei der Entwicklung adaptiver Bildungstechnologien aufgedeckt werden. Zum anderen werden als komplex wahrgenommene Interaktionssituationen identifiziert. An diesen Situationen kann gezeigt werden, wo aufgrund der Komplexität des Systems die etablierten Alltagsroutinen von Nutzenden nicht mehr ausreichen, um die Folgen der Interaktion mit dem System vollständig erfassen zu können. Dieses Wissen kann sowohl Konstrukteur*innen als auch Nutzer*innen helfen, in Zukunft besser mit der inhärenten Komplexität moderner Bildungstechnologien umzugehen.
We introduce a type and effect system, for an imperative object calculus, which infers sharing possibly introduced by the evaluation of an expression, represented as an equivalence relation among its free variables. This direct representation of sharing effects at the syntactic level allows us to express in a natural way, and to generalize, widely-used notions in literature, notably uniqueness and borrowing. Moreover, the calculus is pure in the sense that reduction is defined on language terms only, since they directly encode store. The advantage of this non-standard execution model with respect to a behaviorally equivalent standard model using a global auxiliary structure is that reachability relations among references are partly encoded by scoping. (C) 2018 Elsevier B.V. All rights reserved.
teaspoon
(2018)
Answer Set Programming (ASP) is an approach to declarative problem solving, combining a rich yet simple modeling language with high performance solving capacities. We here develop an ASP-based approach to curriculum-based course timetabling (CB-CTT), one of the most widely studied course timetabling problems. The resulting teaspoon system reads a CB-CTT instance of a standard input format and converts it into a set of ASP facts. In turn, these facts are combined with a first-order encoding for CB-CTT solving, which can subsequently be solved by any off-the-shelf ASP systems. We establish the competitiveness of our approach by empirically contrasting it to the best known bounds obtained so far via dedicated implementations. Furthermore, we extend the teaspoon system to multi-objective course timetabling and consider minimal perturbation problems.
Forschendes Lernen und die digitale Transformation sind zwei der wichtigsten Einflüsse auf die Entwicklung der Hochschuldidaktik im deutschprachigen Raum. Während das forschende Lernen als normative Theorie das sollen beschreibt, geben die digitalen Werkzeuge, alte wie neue, das können in vielen Bereichen vor.
In der vorliegenden Arbeit wird ein Prozessmodell aufgestellt, was den Versuch unternimmt, das forschende Lernen hinsichtlich interaktiver, gruppenbasierter Prozesse zu systematisieren. Basierend auf dem entwickelten Modell wurde ein Softwareprototyp implementiert, der den gesamten Forschungsprozess begleiten kann. Dabei werden Gruppenformation, Feedback- und Reflexionsprozesse und das Peer Assessment mit Bildungstechnologien unterstützt. Die Entwicklungen wurden in einem qualitativen Experiment eingesetzt, um Systemwissen über die Möglichkeiten und Grenzen der digitalen Unterstützung von forschendem Lernen zu gewinnen.
plasp 3
(2019)
We describe the new version of the Planning Domain Definition Language (PDDL)-to-Answer Set Programming (ASP) translator plasp. First, it widens the range of accepted PDDL features. Second, it contains novel planning encodings, some inspired by Satisfiability Testing (SAT) planning and others exploiting ASP features such as well-foundedness. All of them are designed for handling multivalued fluents in order to capture both PDDL as well as SAS planning formats. Third, enabled by multishot ASP solving, it offers advanced planning algorithms also borrowed from SAT planning. As a result, plasp provides us with an ASP-based framework for studying a variety of planning techniques in a uniform setting. Finally, we demonstrate in an empirical analysis that these techniques have a significant impact on the performance of ASP planning.
E-Assessment etablieren
(2020)
Elektronische Lernstandserhebungen, sogenannte E-Assessments, bieten für Lehrende und Studierende viele Vorteile z. B. hinsichtlich schneller Rückmeldungen oder kompetenzorientierter Fragenformate, und ermöglichen es, unabhängig von Ort und Zeit Prüfungen zu absolvieren. In diesem Beitrag werden die Einführung von summativen Lernstandserhebungen, sogenannter E-Klausuren, am Beispiel der Universität Potsdam, der Aufbau einer länderübergreifenden Initiative für E-Assessment sowie technische Möglichkeiten für dezentrale elektronische Klausuren vorgestellt. Dabei werden der aktuelle Stand, die Ziele und die gewählte stufenweise Umsetzungsstrategie der Universität Potsdam skizziert. Darauf aufbauend folgt eine Beschreibung des Vorgehens, der Kooperationsmöglichkeiten für den Wissens- und Erfahrungsaustausch sowie Herausforderungen der E-Assessment- Initiative. Abschließend werden verschiedene E-Klausurformen und technische Möglichkeiten zur Umsetzung komplexer Prüfungsumgebungen klassifiziert sowie mit ihren charakteristischen Vor- und Nachteilen diskutiert und eine integrierte Lösung vorgeschlagen.
Das größte der fächerübergreifenden Projekte im Potsdamer Projekt Qualitätspakt Lehre hatte die flächendeckende Etablierung von digitalen Medien als einen integralen Bestandteil von Lehre und Studium zum Gegenstand. Im Teilprojekt E-Learning in Studienbereichen (eLiS) wurden dafür Maßnahmen in den Feldern Organisations-, technische und Inhaltsentwicklung zusammengeführt. Der vorliegende Beitrag präsentiert auf Basis von Ausgangslage und Zielsetzungen die Ergebnisse rund um die Digitalisierung von Lehre und Studium an der Universität Potsdam. Exemplarisch werden fünf Dienste näher vorgestellt, die inzwischen größtenteils in den Regelbetrieb der Hochschule übergegangen sind: die Videoplattform Media.UP, die mobile App Reflect.UP, die persönliche Lernumgebung Campus. UP, das Self-Service-Portal Cook.UP und das Anzeigesystem Freiraum.UP. Dabei wird jeweils ein technischer Blick „unter die Haube“ verbunden mit einer Erläuterung der Nutzungsmöglichkeiten, denen eine aktuelle Einschätzung von Lehrenden und Studierenden der Hochschule gegenübergestellt wird. Der Beitrag schließt mit einer Einbettung der vorgestellten Entwicklungen in einen größeren Kontext und einem Ausblick auf die weiterhin anstehenden Aufgaben.
Die Setzung strategischer Ziele sowie die Zuordnung und Umsetzung dazugehörender Maßnahmen sind ein wesentliches Element, um die Innovationsfähigkeit von Organisationen zu erhalten. In den vergangenen Jahren ist auch an Hochschulen die Strategiebildung deutlich vorangetrieben worden. Dies betrifft verschiedene Handlungsfelder, und es werden verschiedene Ansätze verfolgt. Der vorliegende Beitrag greift am Beispiel der Universität Potsdam drei in den vergangenen Jahren adressierte Strategiebereiche heraus: IT, E-Learning und Forschungsdaten. Die damit verbundenen Prozesse waren in unterschiedlichem Maß von Partizipation geprägt. Die gesammelten Erfahrungen werden reflektiert, und es werden Empfehlungen für Strategieentwicklungsprozesse abgeleitet.
The Surface Water and Ocean Topography (SWOT) mission is a next generation satellite mission expected to provide a 2 km-resolution observation of the sea surface height (SSH) on a two-dimensional swath. Processing SWOT data will be challenging because of the large amount of data, the mismatch between a high spatial resolution and a low temporal resolution, and the observation errors. The present paper focuses on the reduction of the spatially structured errors of SWOT SSH data. It investigates a new error reduction method and assesses its performance in an observing system simulation experiment. The proposed error-reduction method first projects the SWOT SSH onto a subspace spanned by the SWOT spatially structured errors. This projection is removed from the SWOT SSH to obtain a detrended SSH. The detrended SSH is then processed within an ensemble data assimilation analysis to retrieve a full SSH field. In the latter step, the detrending is applied to both the SWOT data and an ensemble of model-simulated SSH fields. Numerical experiments are performed with synthetic SWOT observations and an ensemble from a North Atlantic, 1/60 degrees simulation of the ocean circulation (NATL60). The data assimilation analysis is carried out with an ensemble Kalman filter. The results are assessed with root mean square errors, power spectrum density, and spatial coherence. They show that a significant part of the large scale SWOT errors is reduced. The filter analysis also reduces the small scale errors and allows for an accurate recovery of the energy of the signal down to 25 km scales. In addition, using the SWOT nadir data to adjust the SSH detrending further reduces the errors.
A common feature in Answer Set Programming is the use of a second negation, stronger than default negation and sometimes called explicit, strong or classical negation. This explicit negation is normally used in front of atoms, rather than allowing its use as a regular operator. In this paper we consider the arbitrary combination of explicit negation with nested expressions, as those defined by Lifschitz, Tang and Turner. We extend the concept of reduct for this new syntax and then prove that it can be captured by an extension of Equilibrium Logic with this second negation. We study some properties of this variant and compare to the already known combination of Equilibrium Logic with Nelson's strong negation.
The target article discusses the question of how educational makerspaces can become places supportive of knowledge construction. This question is too often neglected by people who run makerspaces, as they mostly explain how to use different tools and focus on the creation of a product. In makerspaces, often pupils also engage in physical computing activities and thus in the creation of interactive artifacts containing embedded systems, such as smart shoes or wristbands, plant monitoring systems or drink mixing machines. This offers the opportunity to reflect on teaching physical computing in computer science education, where similarly often the creation of the product is so strongly focused upon that the reflection of the learning process is pushed into the background.
MUP
(2020)
Message Queuing Telemetry Transport (MQTT) is one of the dominating protocols for edge- and cloud-based Internet of Things (IoT) solutions. When a security vulnerability of an IoT device is known, it has to be fixed as soon as possible. This requires a firmware update procedure. In this paper, we propose a secure update protocol for MQTT-connected devices which ensures the freshness of the firmware, authenticates the new firmware and considers constrained devices. We show that the update protocol is easy to integrate in an MQTT-based IoT network using a semantic approach. The feasibility of our approach is demonstrated by a detailed performance analysis of our prototype implementation on a IoT device with 32 kB RAM. Thereby, we identify design issues in MQTT 5 which can help to improve the support of constrained devices.
MUP
(2020)
Message Queuing Telemetry Transport (MQTT) is one of the dominating protocols for edge- and cloud-based Internet of Things (IoT) solutions. When a security vulnerability of an IoT device is known, it has to be fixed as soon as possible. This requires a firmware update procedure. In this paper, we propose a secure update protocol for MQTT-connected devices which ensures the freshness of the firmware, authenticates the new firmware and considers constrained devices. We show that the update protocol is easy to integrate in an MQTT-based IoT network using a semantic approach. The feasibility of our approach is demonstrated by a detailed performance analysis of our prototype implementation on a IoT device with 32 kB RAM. Thereby, we identify design issues in MQTT 5 which can help to improve the support of constrained devices.