Institut für Informatik und Computational Science
Filtern
Volltext vorhanden
- ja (131) (entfernen)
Erscheinungsjahr
Dokumenttyp
- Dissertation (81)
- Wissenschaftlicher Artikel (13)
- Konferenzveröffentlichung (12)
- Masterarbeit (10)
- Postprint (10)
- Monographie/Sammelband (2)
- Bachelorarbeit (1)
- Habilitation (1)
- Bewegte Bilder (1)
Sprache
- Englisch (79)
- Deutsch (51)
- Mehrsprachig (1)
Gehört zur Bibliographie
- ja (131) (entfernen)
Schlagworte
- Maschinelles Lernen (7)
- Antwortmengenprogrammierung (6)
- E-Learning (6)
- Informatik (6)
- Machine Learning (5)
- answer set programming (5)
- Informatikdidaktik (4)
- Modellierung (4)
- Answer Set Programming (3)
- Komplexität (3)
- Ontologie (3)
- Semantic Web (3)
- Softwareentwicklung (3)
- didactics (3)
- education (3)
- higher education (3)
- machine learning (3)
- Algorithmen (2)
- Algorithms (2)
- Assessment (2)
- Autismus (2)
- Bildungstechnologien (2)
- Bildverarbeitung (2)
- Code (2)
- Computergrafik (2)
- Computersicherheit (2)
- Didaktik (2)
- Digitale Medien (2)
- Digitalisierung (2)
- EEG (2)
- Educational Technologies (2)
- FMC (2)
- Hochschuldidaktik (2)
- ICA (2)
- IT-Infrastruktur (2)
- Informatikstudium (2)
- Internet of Things (2)
- Knowledge Representation and Reasoning (2)
- Konstruktivismus (2)
- Künstliche Intelligenz (2)
- Lernumgebung (2)
- MQTT (2)
- Mensch-Technik-Interaktion (2)
- Methodik (2)
- Middleware (2)
- Modell (2)
- Onlinelehre (2)
- Ontology (2)
- Prozessmodellierung (2)
- Software Engineering (2)
- Systemstruktur (2)
- Texturen (2)
- User Experience (2)
- Visualisierung (2)
- Vorhersage (2)
- code (2)
- complexity (2)
- computer graphics (2)
- human computer interaction (2)
- image processing (2)
- informatics (2)
- maschinelles Lernen (2)
- model (2)
- non-photorealistic rendering (2)
- scientific workflows (2)
- software development (2)
- systems biology (2)
- teacher training (2)
- test (2)
- textures (2)
- virtual 3D city models (2)
- virtuelle 3D-Stadtmodelle (2)
- visualization (2)
- 'Peer To Peer' (1)
- 13C metabolic flux analysis (1)
- 3D Computer Grafik (1)
- 3D Computer Graphics (1)
- 3D Drucken (1)
- 3D Linsen (1)
- 3D Semiotik (1)
- 3D Visualisierung (1)
- 3D computer graphics (1)
- 3D lenses (1)
- 3D printing (1)
- 3D semiotics (1)
- 3D visualization (1)
- 3D-Stadtmodelle (1)
- 3d city models (1)
- 6LoWPAN (1)
- ADFS (1)
- ASIC (1)
- ASIC (Applikationsspezifische Integrierte Schaltkreise) (1)
- ASP (Answer Set Programming) (1)
- Abbrecherquote (1)
- Abstraktion (1)
- Accepting Grammars (1)
- Ackerschmalwand (1)
- Active Directory Federation Services (1)
- Active Evaluation (1)
- Adaptivity (1)
- Adaptivität (1)
- Adversarial Learning (1)
- Aktive Evaluierung (1)
- Akzeptierende Grammatiken (1)
- Algorithmenablaufplanung (1)
- Algorithmenkonfiguration (1)
- Algorithmenselektion (1)
- Alignment (1)
- Angewandte Spieltheorie (1)
- Anisotroper Kuwahara Filter (1)
- Anleitung (1)
- Antwortmengen Programmierung (1)
- Application Aggregation (1)
- Applications and Software Development (1)
- Applied Game Theory (1)
- Arabidopsis thaliana (1)
- Argumentation (1)
- Artificial Intelligence (1)
- Asynchrone Schaltung (1)
- Attention (1)
- Aufmerksamkeit (1)
- Aufzählung (1)
- Augenbewegungen (1)
- Ausbildung (1)
- Ausreissererkennung (1)
- Authorization (1)
- Autism (1)
- Automatic UI Generation (1)
- Autorisierung (1)
- BCH (1)
- BCH code (1)
- BCH-Code (1)
- BCI (1)
- BSS (1)
- Bachelorstudierende der Informatik (1)
- Baumweite (1)
- Behavior (1)
- Benutzeroberfläche (1)
- Benutzungsschnittstellen Ontologien (1)
- Berührungseingaben (1)
- Betrachtungsebenen (1)
- Beweis (1)
- Beweisassistent (1)
- Beweistheorie (1)
- Beweisumgebung (1)
- Bilddatenanalyse (1)
- Bildung (1)
- Binäres Entscheidungsdiagramm (1)
- Bio-jETI (1)
- Biocomputing (1)
- Bioelektrisches Signal (1)
- Bioinformatik (1)
- Boolean constraint solver (1)
- Boosting (1)
- Brain Computer Interface (1)
- Business Process Models (1)
- CASP (Constraint Answer Set Programming) (1)
- CSC (1)
- Cactus (1)
- Choreographien (1)
- CityGML (1)
- Classification (1)
- Clusteranalyse (1)
- Codierung (1)
- Codierungstheorie (1)
- Coding theory (1)
- Cognitive Apprenticeship (1)
- Common Spatial Pattern (1)
- Complementary Circuits (1)
- Complexity (1)
- Compliance (1)
- Composed UIs (1)
- Composition (1)
- Computational Complexity (1)
- Computational Science (1)
- Computer Science (1)
- Computer Science Education (1)
- Computergestützes Training (1)
- Conceptual (1)
- Constraint (1)
- Constraint Solving (1)
- Constraint-Programmierung (1)
- Constructive solid geometry (1)
- Controlled Derivations (1)
- Coq (1)
- Covariate Shift (1)
- Curry (1)
- DDoS (1)
- DNA (1)
- DNA computing (1)
- DNS (1)
- DPLL (1)
- Debugging (1)
- Deduction (1)
- Deep Learning (1)
- Defining characteristics of physical computing (1)
- Dempster-Shafer-Theorie (1)
- Dempster–Shafer theory (1)
- Description Logics (1)
- Deskriptive Logik (1)
- Diagonalisierung (1)
- Dialog-based User Interfaces (1)
- Dialogbasierte Benutzerschnittstellen (1)
- Didaktik der Informatik (1)
- Didaktische Konzepte (1)
- Dienst-Ökosysteme (1)
- Dienstkomposition (1)
- Dienstplattform (1)
- Differenz von Gauss Filtern (1)
- Digital Design (1)
- Digital Game Based Learning (1)
- Digital Media (1)
- Digitalisation (1)
- Distributed Computing (1)
- Diversität (1)
- Domain-Specific Languages (1)
- Domänenspezifische Sprachen (1)
- Dreidimensionale Computergraphik (1)
- Dynamic Programming (1)
- Dynamische Programmierung (1)
- Dynamische Rekonfiguration (1)
- E-Assessment (1)
- E-Government (1)
- E-Klausuren (1)
- E-Portfolio (1)
- Echtzeitanwendung (1)
- Eingabegenauigkeit (1)
- Eisenbahnnetz (1)
- Elektroencephalographie (1)
- Elektronisches Prüfen (1)
- Emotionen (1)
- Emotionsforschung (1)
- Enterprise Search (1)
- Entscheidungsbäume (1)
- Entwurf (1)
- Entwurfsmuster (1)
- Entwurfsmuster für SOA-Sicherheit (1)
- Entwurfsprinzipien (1)
- Entwurfsraumexploration (1)
- Equilibrium logic (1)
- Erfüllbarkeit einer Formel der Aussagenlogik (1)
- Erfüllbarkeitsproblem (1)
- Error Estimation (1)
- Error-Detection Circuits (1)
- European Bioinformatics Institute (1)
- Evaluierung semantischer Suchmaschinen (1)
- Evidenztheorie (1)
- Exploration (1)
- Exponential Time Hypothesis (1)
- Exponentialzeit Hypothese (1)
- FMC-QE (1)
- FOSS (1)
- FPGA (1)
- Feature Combination (1)
- Feedback (1)
- Fehlende Daten (1)
- Fehlererkennung (1)
- Fehlerkorrektur (1)
- Fehlerschätzung (1)
- Fehlvorstellung (1)
- Flussgesteuerter Bilateraler Filter (1)
- Focus+Context Visualization (1)
- Fokus-&-Kontext Visualisierung (1)
- Formalismus (1)
- Formalitätsgrad (1)
- Formeln der quantifizierten Aussagenlogik (1)
- Forschendes Lernen (1)
- Forschungsdatenmanagement (1)
- GIS-Dienstkomposition (1)
- GPU (1)
- Game-Design-Elemente (1)
- Game-based learning (1)
- Gamification (1)
- Gebrauchstauglichkeit (1)
- Gebäudemodelle (1)
- Gehirn-Computer-Schnittstelle (1)
- Geländemodelle (1)
- Generalisierung (1)
- Generative Programmierung (1)
- Generative Programming (1)
- Geodaten (1)
- Geometrieerzeugung (1)
- Geovisualisierung (1)
- Geschäftsprozessmodelle (1)
- Gesichtsausdruck (1)
- Gesteuerte Ableitungen (1)
- Gleichheit (1)
- Globus (1)
- Grammar Systems (1)
- Grammatikalische Inferenz (1)
- Grammatiksysteme (1)
- Graph-basiertes Ranking (1)
- Graphfärbung (1)
- Grid (1)
- Grid Computing (1)
- HCI (1)
- HDI (1)
- Hardware Design (1)
- Hardware-Software-Co-Design (1)
- Hauptkomponentenanalyse (1)
- Heterogenität (1)
- High-Level Synthesis (1)
- Hochschulbildung (1)
- Hochschullehre (1)
- Hochschulsystem (1)
- Human-Technology Interaction (1)
- I/O-effiziente Algorithmen (1)
- ICT (1)
- IP core (1)
- IT security (1)
- IT-Security (1)
- IT-Sicherheit (1)
- Informatics (1)
- Informatik-Studiengänge (1)
- Informatiksystem (1)
- Informatikunterricht (1)
- Informatikvoraussetzungen (1)
- Information Transfer Rate (1)
- Informationsextraktion (1)
- Infrastruktur (1)
- Inkonsistenz (1)
- Innovation (1)
- Inquiry-based learning (1)
- Integration (1)
- Interactive Rendering (1)
- Interactive system (1)
- Interaktionsmodel (1)
- Interaktionsmodellierung (1)
- Interaktionstechniken (1)
- Interaktives Rendering (1)
- Interaktives System (1)
- Interface design (1)
- Internet Security (1)
- Internet-Sicherheit (1)
- Interoperability (1)
- Interoperabilität (1)
- Intuition (1)
- IoT (1)
- Kartografisches Design (1)
- Kern-PCA (1)
- Kernmethoden (1)
- Key Competencies (1)
- Klassifikation (1)
- Klassifikation mit großem Margin (1)
- Klassifikator-Kalibrierung (1)
- Klimafolgenanalyse (1)
- Klimawandel (1)
- Kommunikation (1)
- Kompetenz (1)
- Komplexitätsbewältigung (1)
- Komplexitätstheorie (1)
- Komposition (1)
- Konnektionskalkül (1)
- Konzeptionell (1)
- Kybernetik (1)
- LDPC code (1)
- LDPC-Code (1)
- Landmarken (1)
- Large Margin Classification (1)
- Laser Cutten (1)
- Learning Analytics (1)
- Learning environment (1)
- Lebenslanges Lernen (1)
- Leftmost Derivations (1)
- Lehrer (1)
- Lehrkräfteausbildung (1)
- Leistungsvorhersage (1)
- Lernsoftware (1)
- Life-Long Learning (1)
- Linked Data Anwendungen (1)
- Linked Data Application Modelling (1)
- Linksableitungen (1)
- Logic Programming (1)
- Logics (1)
- Logik (1)
- Logiksynthese (1)
- Lower Bounds (1)
- MEG (1)
- MFA (1)
- Magnetoencephalographie (1)
- Malware (1)
- Mathematical Optimization (1)
- Mathematikdidaktik (1)
- Mathematikphilosophie (1)
- Mathematische Optimierung (1)
- Matrizen-Eigenwertaufgabe (1)
- Megamodel (1)
- Megamodell (1)
- Mehrklassen-Klassifikation (1)
- Mensch-Computer-Interaktion (1)
- Message Passing Interface (1)
- Metamodell (1)
- Methoden der semantischen Suche (1)
- Methodology (1)
- Migration (1)
- Mischmodelle (1)
- Mischung <Signalverarbeitung> (1)
- Mobile App (1)
- Mobiles Lernen (1)
- Mobilgeräte (1)
- Model Based Engineering (1)
- Model Checking (1)
- Model Driven Architecture (1)
- Model Driven UI Development (1)
- Model Management (1)
- Model-Driven Engineering (1)
- Model-Driven Software Development (1)
- Modeling (1)
- Modell Management (1)
- Modell-driven Security (1)
- Modell-getriebene Sicherheit (1)
- Modellbasiert (1)
- Modellgetriebene Architektur (1)
- Modellgetriebene Entwicklung (1)
- Modellgetriebene Softwareentwicklung (1)
- Modellgetriebene UI Entwicklung (1)
- Modelling (1)
- Molekulare Bioinformatik (1)
- Motivation (1)
- Multi Task Learning (1)
- Multi-Class (1)
- Multi-Task-Lernen (1)
- Multimodal User Interfaces (1)
- Multimodale Benutzerschnittstellen (1)
- Multiprocessor (1)
- Multiprozessor (1)
- NETCONF (1)
- NUI (1)
- Navigation (1)
- Network Management (1)
- Netzwerk Management (1)
- Netzwerke (1)
- Neuronales Netz (1)
- New On-Line Error-Detection Methode (1)
- Next Generation Network (1)
- Nicht-photorealistisches Rendering (1)
- Nichtfotorealistische Bildsynthese (1)
- Nutzungserlebnis (1)
- Nutzungsinteresse (1)
- Objektive Schwierigkeit (1)
- Omega (1)
- Ontologien (1)
- Ontologies (1)
- Open Source (1)
- Optimierung (1)
- Optimierungsproblem (1)
- Optimization (1)
- Parallel Programming (1)
- Parallele Datenverarbeitung (1)
- Paralleles Rechnen (1)
- Parallelrechner (1)
- Parameterized Complexity (1)
- Parametrisierte Komplexität (1)
- Parsing (1)
- Partizipation (1)
- Patterns (1)
- Peer-to-Peer-Netz ; GRID computing ; Zuverlässigkeit ; Web Services ; Betriebsmittelverwaltung ; Migration (1)
- Performance Prediction (1)
- Persönliche Lernumgebung (1)
- Planing (1)
- Platzierung (1)
- Policy Enforcement (1)
- Power Monitoring (1)
- Pre-RS Traceability (1)
- Prediction Game (1)
- Predictive Models (1)
- Probleme in der Studie (1)
- Problemlösen (1)
- Process (1)
- Process modeling (1)
- Professoren (1)
- Programmierung (1)
- Proof Theory (1)
- Prototyp (1)
- Prototyping (1)
- Prozess (1)
- Prozess Verbesserung (1)
- Prozesse (1)
- Prozessmodell (1)
- Prozesssynchronisierung (1)
- Prädiktionsspiel (1)
- Präferenzen (1)
- Quantified Boolean Formula (QBF) (1)
- Quantitative Modeling (1)
- Quantitative Modellierung (1)
- Queuing Theory (1)
- Reconfigurable (1)
- Regression (1)
- Regularisierung (1)
- Regularization (1)
- Rekonfiguration (1)
- Rendering (1)
- Reparatur (1)
- Reuseable UIs (1)
- SMT (SAT Modulo Theories) (1)
- SOA Security Pattern (1)
- SSO (1)
- STG decomposition (1)
- STG-Dekomposition (1)
- Sample Selection Bias (1)
- Satisfiability (1)
- Scene graph systems (1)
- Schlüsselkompetenzen (1)
- Schulmaterial (1)
- Security Modelling (1)
- Segmentierung (1)
- Selektionsbias (1)
- Self-Checking Circuits (1)
- Semantic Search (1)
- Semantik Web (1)
- Semantische Suche (1)
- Sensornetzwerke (1)
- Service Creation (1)
- Service Delivery Platform (1)
- Service Ecosystems (1)
- Service Oriented Architectures (1)
- Service convergence (1)
- Service-Orientierte Architekturen (1)
- Service-oriented Architectures (1)
- Serviceorientierte Architektur (1)
- Shader (1)
- Sicherheitsmodellierung (1)
- Signal Processing (1)
- Signalquellentrennung (1)
- Signaltrennung (1)
- Simultane Diagonalisierung (1)
- Single Event Transient (1)
- Single Sign On (1)
- Single Trial Analysis (1)
- Skelettberechnung (1)
- Software (1)
- Software architecture (1)
- Software-basierte Cache-Kohärenz (1)
- Softwarearchitektur (1)
- Sonnenteilchen-Ereignis (1)
- Spam (1)
- Spam Filtering (1)
- Spam-Erkennung (1)
- Spam-Filter (1)
- Spam-Filtering (1)
- Spatio-Spectral Filter (1)
- Spawning (1)
- Spielbasiertes Lernen (1)
- Sprachdesign (1)
- Statistical Tests (1)
- Statistikprogramm R (1)
- Statistische Tests (1)
- Stilisierung (1)
- Strahlungshartes Design (1)
- Strahlungshärte Entwurf (1)
- Strategie (1)
- Stromverbrauchüberwachung (1)
- Structuring (1)
- Strukturierung (1)
- Studentenerwartungen (1)
- Studentenhaltungen (1)
- Studentenjobs (1)
- Studienabbrecher (1)
- Studiendauer (1)
- Suche (1)
- Support Vectors (1)
- Support-Vector Lernen (1)
- Synthese (1)
- System Biologie (1)
- System structure (1)
- Systembiologie (1)
- Systementwurf (1)
- Szenengraph (1)
- TPTP (1)
- Tailored UI Variants (1)
- Taktik (1)
- Telekommunikation (1)
- Temporal Logic (1)
- Temporallogik (1)
- Temporäre Anbindung (1)
- Terminologische Logik (1)
- Test (1)
- Theoretische Informatik (1)
- Theoretischen Vorlesungen (1)
- Time Augmented Petri Nets (1)
- Time Series Analysis (1)
- Tool (1)
- Traceability (1)
- Tracking (1)
- Transformation (1)
- Treewidth (1)
- UI Components (1)
- UI Metamodels (1)
- UI-Komponenten (1)
- Unabhängige Komponentenanalyse (1)
- Universität Bagdad (1)
- Universität Potsdam (1)
- Universitätseinstellungen (1)
- Untere Schranken (1)
- Unterrichtswerkzeuge (1)
- Unvollständigkeit (1)
- Usability (1)
- Usage Interest (1)
- User Interface Ontologies (1)
- User Interfaces (1)
- VM (1)
- Verhalten (1)
- Verifikation (1)
- Verletzung Auflösung (1)
- Verletzung Erklärung (1)
- Verteiltes Rechnen (1)
- Verteilungsunterschied (1)
- Violation Explanation (1)
- Violation Resolution (1)
- Virtual Reality (1)
- Virtuelles 3D Stadtmodell (1)
- Vorhersagemodelle (1)
- Wahrnehmung (1)
- Wahrnehmung von Arousal (1)
- Wahrnehmungsunterschiede (1)
- Warteschlangentheorie (1)
- Web Services (1)
- Web Sites (1)
- Web of Data (1)
- Webseite (1)
- Well-structuredness (1)
- Wetterextreme (1)
- Wirtschaftsinformatik (1)
- Wissenschaftlichesworkflows (1)
- Wissensrepräsentation und -verarbeitung (1)
- Wissensrepräsentation und Schlussfolgerung (1)
- Wohlstrukturiertheit (1)
- Workflow (1)
- ZQSA (1)
- ZQSAT (1)
- Zeitbehaftete Petri Netze (1)
- Zero-Suppressed Binary Decision Diagram (ZDD) (1)
- Zuverlässigkeitsanalyse (1)
- abstraction (1)
- adaptiv (1)
- adaptive (1)
- algorithm configuration (1)
- algorithm scheduling (1)
- algorithm selection (1)
- animated PCA (1)
- animierte PCA (1)
- anisotropic Kuwahara filter (1)
- anti-cancer drugs (1)
- approximate joint diagonalization (1)
- argument mining (1)
- argumentation (1)
- argumentation structure (1)
- arithmethische Prozeduren (1)
- arithmetic procedures (1)
- arousal perception (1)
- artificial intelligence (1)
- asynchronous circuit (1)
- automatic feedback (1)
- automatic theorem prover (1)
- automatisierter Theorembeweiser (1)
- bild (1)
- bio-computing (1)
- biometrics (1)
- biometrische Identifikation (1)
- blind source separation (1)
- building models (1)
- business informatics (1)
- cartographic design (1)
- changing the study field (1)
- changing the university (1)
- choreographies (1)
- classifier calibration (1)
- classroom material (1)
- climate change (1)
- climate impact analysis (1)
- clustering (1)
- cognitive apprenticeship (1)
- coherence-enhancing filtering (1)
- communication (1)
- computational biology (1)
- computational methods (1)
- computational thinking (1)
- computer science education (1)
- computer security (1)
- computergestützte Methoden (1)
- concurrent checking (1)
- connection calculus (1)
- constraint (1)
- constraint programming (1)
- constraints (1)
- constructivism (1)
- construktivism (1)
- debugging (1)
- decision trees (1)
- deep neural networks (1)
- degree of formality (1)
- design (1)
- design principles (1)
- design space exploration (1)
- didaktische Rekonstruktion (1)
- difference of Gaussians (1)
- digital circuit (1)
- digital design (1)
- digitale Hochschullehre (1)
- digitally-enabled pedagogies (1)
- dropout (1)
- drug-sensitivity prediction (1)
- dynamic (1)
- dynamic classification (1)
- dynamic reconfiguration (1)
- dynamisch (1)
- dynamische Klassifikation (1)
- e-Learning (1)
- eGovernment (1)
- edge computing (1)
- educational reconstruction (1)
- educational systems (1)
- eingebettete Systeme (1)
- einseitige Kommunikation (1)
- email spam detection (1)
- embedded systems (1)
- emotion (1)
- emotion representation (1)
- emotion research (1)
- enterprise search (1)
- entity alignment (1)
- enumeration (1)
- epistemic logic programs (1)
- epistemic specifications (1)
- equality (1)
- error correction (1)
- error detection (1)
- evidence theory (1)
- explicit negation (1)
- external memory algorithms (1)
- eye movements (1)
- face tracking (1)
- facial expression (1)
- firmware update (1)
- flow-based bilateral filter (1)
- formalism (1)
- freie Daten (1)
- freie Software (1)
- game based learning (1)
- game design elements (1)
- game-based learning (1)
- generalization (1)
- geometry generation (1)
- geospatial data (1)
- geospatial services (1)
- geovisualization (1)
- global constraints (1)
- globale Constraints (1)
- grammar inference (1)
- graph clustering (1)
- graph-based ranking (1)
- hardware design (1)
- hardware-software-codesign (1)
- high-throughput analysis (1)
- human-technology interaction (1)
- hybrid (1)
- hybrid semantic search (1)
- hybride semantische Suche (1)
- hybrides Problemlösen (1)
- image (1)
- image data analysis (1)
- incompleteness (1)
- inconsistency (1)
- independent component analysis (1)
- indirect economic impacts (1)
- indirekte ökonomische Effekte (1)
- informal and formal learning (1)
- information extraction (1)
- information retrieval (1)
- informatische Bildung im Sekundarbereich (1)
- infrastructure (1)
- input accuracy (1)
- interaction modeling (1)
- interaction techniques (1)
- intuition (1)
- kernel PCA (1)
- kernel methods (1)
- key competences in physical computing (1)
- konvergente Dienste (1)
- landmarks (1)
- language design (1)
- lautes Denken (1)
- leanCoP (1)
- learning environment (1)
- lebenslanges Lernen (1)
- lesson planning (1)
- lesson preparation (1)
- linear code (1)
- linearer Code (1)
- logic (1)
- logic programming (1)
- logic synthesis (1)
- logical errors (1)
- logical signaling networks (1)
- logische Ergänzung (1)
- logische Fehler (1)
- logische Programmierung (1)
- logische Signalnetzwerke (1)
- macro-economic modelling (1)
- makroökonomische Modellierung (1)
- malware detection (1)
- map/reduce (1)
- maschninelles Lernen (1)
- mathematics education (1)
- medical (1)
- medizinisch (1)
- meta model (1)
- methodology (1)
- middleware (1)
- misconception (1)
- mixture models (1)
- mobile devices (1)
- mobile learning (1)
- mobile technologies and apps (1)
- mobiles lernen (1)
- model-based (1)
- model-driven architecture (1)
- modeling (1)
- molecular networks (1)
- molekulare Netzwerke (1)
- multi core data processing (1)
- multi-class classification (1)
- navigation (1)
- networks-on-chip (1)
- neue Online-Fehlererkennungsmethode (1)
- nichtlineare ICA (1)
- nichtlineare PCA (NLPCA) (1)
- nichtlineare Projektionen (1)
- non-monotonic reasoning (1)
- nonlinear ICA (1)
- nonlinear PCA (NLPCA) (1)
- nonlinear projections (1)
- objective difficulty (1)
- omega (1)
- on-chip (1)
- one-sided communication (1)
- oneM2M (1)
- ontologies (1)
- open source (1)
- optimization (1)
- organisational evolution (1)
- outlier detection (1)
- output space compaction (1)
- overcomplete ICA (1)
- paper prototyping (1)
- parallel programming (1)
- parallel solving (1)
- parallele Programmierung (1)
- paralleles Lösen (1)
- parsing (1)
- pattern recognition (1)
- perception (1)
- perception differences (1)
- philosophy of mathematics (1)
- physical Computing (1)
- physical computing (1)
- physical computing tools (1)
- placement (1)
- policy evaluation (1)
- prediction (1)
- preferences (1)
- priorities (1)
- probabilistic deep learning (1)
- probabilistic deep metric learning (1)
- probabilistische tiefe neuronale Netze (1)
- probabilistisches tiefes metrisches Lernen (1)
- process (1)
- process improvement (1)
- process model (1)
- process modelling (1)
- process synchronization (1)
- professors (1)
- proof (1)
- proof assistant (1)
- proof environment (1)
- propagation probability (1)
- prototype (1)
- radiation hardness (1)
- radiation hardness design (1)
- railway network (1)
- real-time application (1)
- reconfiguration (1)
- rekonfigurierbar (1)
- reliability assessment (1)
- repair (1)
- robust ICA (1)
- robuste ICA (1)
- scheduling (1)
- search (1)
- secondary computer science education (1)
- security (1)
- segmentation (1)
- selbstanpassendes Multiprozessorsystem (1)
- selbstprüfende Schaltungen (1)
- self-adaptive multiprocessing system (1)
- semantic domain modeling (1)
- semantic ranking (1)
- semantic search (1)
- semantic search evaluation (1)
- semantic search methods (1)
- semantische Domänenmodellierung (1)
- semantische Suche (1)
- semantisches Netz (1)
- semantisches Ranking (1)
- service composition (1)
- shader (1)
- single event upset (1)
- skeletonization (1)
- socio-technical system (1)
- software (1)
- software engineering (1)
- software-based cache coherence (1)
- solar particle event (1)
- sozio-technisches System (1)
- speed independence (1)
- statistics program R (1)
- strahleninduzierte Einzelereignis-Effekte (1)
- structured output prediction (1)
- strukturierte Vorhersage (1)
- study problems (1)
- stylization (1)
- support system (1)
- systematic (1)
- systematisch (1)
- tactic (1)
- teachers (1)
- temporary binding (1)
- terrain models (1)
- think aloud (1)
- tools for teaching (1)
- topics (1)
- touch input (1)
- tptp (1)
- transformation (1)
- tutorial section (1)
- university education (1)
- verification (1)
- virtual 3D city model (1)
- virtual machine (1)
- virtual mobility (1)
- weather extremes (1)
- workflow management (1)
- zero-aliasing (1)
- überbestimmte ICA (1)
Institut
- Institut für Informatik und Computational Science (131)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (17)
- Extern (4)
- eLiS - E-Learning in Studienbereichen (2)
- Historisches Institut (1)
- Potsdam Institute for Climate Impact Research (PIK) e. V. (1)
- Universitätsbibliothek (1)
- Wirtschaftswissenschaften (1)
- Zentrum für Qualitätsentwicklung in Lehre und Studium (ZfQ) (1)
Accurately solving classification problems nowadays is likely to be the most relevant machine learning task. Binary classification separating two classes only is algorithmically simpler but has fewer potential applications as many real-world problems are multi-class. On the reverse, separating only a subset of classes simplifies the classification task. Even though existing multi-class machine learning algorithms are very flexible regarding the number of classes, they assume that the target set Y is fixed and cannot be restricted once the training is finished. On the other hand, existing state-of-the-art production environments are becoming increasingly interconnected with the advance of Industry 4.0 and related technologies such that additional information can simplify the respective classification problems. In light of this, the main aim of this thesis is to introduce dynamic classification that generalizes multi-class classification such that the target class set can be restricted arbitrarily to a non-empty class subset M of Y at any time between two consecutive predictions.
This task is solved by a combination of two algorithmic approaches. First, classifier calibration, which transforms predictions into posterior probability estimates that are intended to be well calibrated. The analysis provided focuses on monotonic calibration and in particular corrects wrong statements that appeared in the literature. It also reveals that bin-based evaluation metrics, which became popular in recent years, are unjustified and should not be used at all. Next, the validity of Platt scaling, which is the most relevant parametric calibration approach, is analyzed in depth. In particular, its optimality for classifier predictions distributed according to four different families of probability distributions as well its equivalence with Beta calibration up to a sigmoidal preprocessing are proven. For non-monotonic calibration, extended variants on kernel density estimation and the ensemble method EKDE are introduced. Finally, the calibration techniques are evaluated using a simulation study with complete information as well as on a selection of 46 real-world data sets.
Building on this, classifier calibration is applied as part of decomposition-based classification that aims to reduce multi-class problems to simpler (usually binary) prediction tasks. For the involved fusing step performed at prediction time, a new approach based on evidence theory is presented that uses classifier calibration to model mass functions. This allows the analysis of decomposition-based classification against a strictly formal background and to prove closed-form equations for the overall combinations. Furthermore, the same formalism leads to a consistent integration of dynamic class information, yielding a theoretically justified and computationally tractable dynamic classification model. The insights gained from this modeling are combined with pairwise coupling, which is one of the most relevant reduction-based classification approaches, such that all individual predictions are combined with a weight. This not only generalizes existing works on pairwise coupling but also enables the integration of dynamic class information.
Lastly, a thorough empirical study is performed that compares all newly introduced approaches to existing state-of-the-art techniques. For this, evaluation metrics for dynamic classification are introduced that depend on corresponding sampling strategies. Thereafter, these are applied during a three-part evaluation. First, support vector machines and random forests are applied on 26 data sets from the UCI Machine Learning Repository. Second, two state-of-the-art deep neural networks are evaluated on five benchmark data sets from a relatively recent reference work. Here, computationally feasible strategies to apply the presented algorithms in combination with large-scale models are particularly relevant because a naive application is computationally intractable. Finally, reference data from a real-world process allowing the inclusion of dynamic class information are collected and evaluated. The results show that in combination with support vector machines and random forests, pairwise coupling approaches yield the best results, while in combination with deep neural networks, differences between the different approaches are mostly small to negligible. Most importantly, all results empirically confirm that dynamic classification succeeds in improving the respective prediction accuracies. Therefore, it is crucial to pass dynamic class information in respective applications, which requires an appropriate digital infrastructure.
As a result of CMOS scaling, radiation-induced Single-Event Effects (SEEs) in electronic circuits became a critical reliability issue for modern Integrated Circuits (ICs) operating under harsh radiation conditions. SEEs can be triggered in combinational or sequential logic by the impact of high-energy particles, leading to destructive or non-destructive faults, resulting in data corruption or even system failure. Typically, the SEE mitigation methods are deployed statically in processing architectures based on the worst-case radiation conditions, which is most of the time unnecessary and results in a resource overhead. Moreover, the space radiation conditions are dynamically changing, especially during Solar Particle Events (SPEs). The intensity of space radiation can differ over five orders of magnitude within a few hours or days, resulting in several orders of magnitude fault probability variation in ICs during SPEs. This thesis introduces a comprehensive approach for designing a self-adaptive fault resilient multiprocessing system to overcome the static mitigation overhead issue. This work mainly addresses the following topics: (1) Design of on-chip radiation particle monitor for real-time radiation environment detection, (2) Investigation of space environment predictor, as support for solar particle events forecast, (3) Dynamic mode configuration in the resilient multiprocessing system. Therefore, according to detected and predicted in-flight space radiation conditions, the target system can be configured to use no mitigation or low-overhead mitigation during non-critical periods of time. The redundant resources can be used to improve system performance or save power. On the other hand, during increased radiation activity periods, such as SPEs, the mitigation methods can be dynamically configured appropriately depending on the real-time space radiation environment, resulting in higher system reliability. Thus, a dynamic trade-off in the target system between reliability, performance and power consumption in real-time can be achieved. All results of this work are evaluated in a highly reliable quad-core multiprocessing system that allows the self-adaptive setting of optimal radiation mitigation mechanisms during run-time. Proposed methods can serve as a basis for establishing a comprehensive self-adaptive resilient system design process. Successful implementation of the proposed design in the quad-core multiprocessor shows its application perspective also in the other designs.
Contemporary multi-core processors are parallel systems that also provide shared memory for programs running on them. Both the increasing number of cores in so-called many-core systems and the still growing computational power of the cores demand for memory systems that are able to deliver high bandwidths. Caches are essential components to satisfy this requirement. Nevertheless, hardware-based cache coherence in many-core chips faces practical limits to provide both coherence and high memory bandwidths. In addition, a shift away from global coherence can be observed. As a result, alternative architectures and suitable programming models need to be investigated.
This thesis focuses on fast communication for non-cache-coherent many-core architectures. Experiments are conducted on the Single-Chip Cloud Computer (SCC), a non-cache-coherent many-core processor with 48 mesh-connected cores. Although originally designed for message passing, the results of this thesis show that shared memory can be efficiently used for one-sided communication on this kind of architecture. One-sided communication enables data exchanges between processes where the receiver is not required to know the details of the performed communication. In the notion of the Message Passing Interface (MPI) standard, this type of communication allows to access memory of remote processes. In order to support this communication scheme on non-cache-coherent architectures, both an efficient process synchronization and a communication scheme with software-managed cache coherence are designed and investigated.
The process synchronization realizes the concept of the general active target synchronization scheme from the MPI standard. An existing classification of implementation approaches is extended and used to identify an appropriate class for the non-cache-coherent shared memory platform. Based on this classification, existing implementations are surveyed in order to find beneficial concepts, which are then used to design a lightweight synchronization protocol for the SCC that uses shared memory and uncached memory accesses. The proposed scheme is not prone to process skew and also enables direct communication as soon as both communication partners are ready. Experimental results show very good scaling properties and up to five times lower synchronization latency compared to a tuned message-based MPI implementation for the SCC.
For the communication, SCOSCo, a shared memory approach with software-managed cache coherence, is presented. According requirements for the coherence that fulfill MPI's separate memory model are formulated, and a lightweight implementation exploiting SCC hard- and software features is developed. Despite a discovered malfunction in the SCC's memory subsystem, the experimental evaluation of the design reveals up to five times better bandwidths and nearly four times lower latencies in micro-benchmarks compared to the SCC-tuned but message-based MPI library. For application benchmarks, like a parallel 3D fast Fourier transform, the runtime share of communication can be reduced by a factor of up to five. In addition, this thesis postulates beneficial hardware concepts that would support software-managed coherence for one-sided communication on future non-cache-coherent architectures where coherence might be only available in local subdomains but not on a global processor level.
Lehrende in der Lehrkräfteausbildung sind stets damit konfrontiert, dass sie den Studierenden innovative Methoden modernen Schulunterrichts traditionell rezipierend vorstellen. In Deutschland gibt es circa 40 Universitäten, die Informatik mit Lehramtsbezug ausbilden. Allerdings gibt es nur wenige Konzepte, die sich mit der Verbindung von Bildungswissenschaften und der Informatik mit ihrer Didaktik beschäftigen und keine Konzepte, die eine konstruktivistische Lehre in der Informatik verfolgen.
Daher zielt diese Masterarbeit darauf ab, diese Lücke aufgreifen und anhand des „Didaktik der Informatik I“ Moduls der Universität Potsdam ein Modell zur konstruktivistischen Hochschullehre zu entwickeln. Dabei soll ein bestehendes konstruktivistisches Lehrmodell auf die Informatikdidaktik übertragen und Elemente zur Verbindung von Bildungswissenschaften, Fachwissenschaften und Fachdidaktiken mit einbezogen werden. Dies kann eine Grundlage für die Planung von Informatikdidaktischen Modulen bieten, aber auch als Inspiration zur Übertragung bestehender innovativer Lehrkonzepte auf andere Fachdidaktiken dienen.
Um ein solches konstruktivistisches Lehr-Lern-Modell zu erstellen, wird zunächst der Zusammenhang von Bildungswissenschaften, Fachwissenschaften und Fachdidaktiken erläutert und anschließend die Notwendigkeit einer Vernetzung hervorgehoben. Hieran folgt eine Darstellung zu relevanten Lerntheorien und bereits entwickelten innovativen Lernkonzepten. Anknüpfend wird darauf eingegangen, welche Anforderungen die Kultusminister- Konferenz an die Ausbildung von Lehrkräften stellt und wie diese Ausbildung für die Informatik momentan an der Universität Potsdam erfolgt. Aus allen Erkenntnissen heraus werden Anforderungen an ein konstruktivistisches Lehrmodell festgelegt. Unter Berücksichtigung der Voraussetzungen der Studienordnung für das Lehramt Informatik wird anschließend ein Modell für konstruktivistische Informatikdidaktik vorgestellt.
Weiterführende Forschung könnte sich damit auseinandersetzen, inwiefern sich die Motivation und Leistung im vergleich zum ursprünglichen Modul ändert und ob die Kompetenzen zur Unterrichtsplanung und Unterrichtsgestaltung durch das neue Modulkonzept stärker ausgebaut werden können.
KEYCIT 2014
(2015)
In our rapidly changing world it is increasingly important not only to be an expert in a chosen field of study but also to be able to respond to developments, master new approaches to solving problems, and fulfil changing requirements in the modern world and in the job market. In response to these needs key competencies in understanding, developing and using new digital technologies are being brought into focus in school and university programmes. The IFIP TC3 conference "KEYCIT – Key Competences in Informatics and ICT (KEYCIT 2014)" was held at the University of Potsdam in Germany from July 1st to 4th, 2014 and addressed the combination of key competencies, Informatics and ICT in detail. The conference was organized into strands focusing on secondary education, university education and teacher education (organized by IFIP WGs 3.1 and 3.3) and provided a forum to present and to discuss research, case studies, positions, and national perspectives in this field.
With the rise of electronic integration between organizations, the need for a precise specification of interaction behavior increases. Information systems, replacing interaction previously carried out by humans via phone, faxes and emails, require a precise specification for handling all possible situations. Such interaction behavior is described in process choreographies. Choreographies enumerate the roles involved, the allowed interactions, the message contents and the behavioral dependencies between interactions. Choreographies serve as interaction contract and are the starting point for adapting existing business processes and systems or for implementing new software components. As a thorough analysis and comparison of choreography modeling languages is missing in the literature, this thesis introduces a requirements framework for choreography languages and uses it for comparing current choreography languages. Language proposals for overcoming the limitations are given for choreography modeling on the conceptual and on the technical level. Using an interconnection modeling style, behavioral dependencies are defined on a per-role basis and different roles are interconnected using message flow. This thesis reveals a number of modeling "anti-patterns" for interconnection modeling, motivating further investigations on choreography languages following the interaction modeling style. Here, interactions are seen as atomic building blocks and the behavioral dependencies between them are defined globally. Two novel language proposals are put forward for this modeling style which have already influenced industrial standardization initiatives. While avoiding many of the pitfalls of interconnection modeling, new anomalies can arise in interaction models. A choreography might not be realizable, i.e. there does not exist a set of interacting roles that collectively realize the specified behavior. This thesis investigates different dimensions of realizability.
Forschendes Lernen und die digitale Transformation sind zwei der wichtigsten Einflüsse auf die Entwicklung der Hochschuldidaktik im deutschprachigen Raum. Während das forschende Lernen als normative Theorie das sollen beschreibt, geben die digitalen Werkzeuge, alte wie neue, das können in vielen Bereichen vor.
In der vorliegenden Arbeit wird ein Prozessmodell aufgestellt, was den Versuch unternimmt, das forschende Lernen hinsichtlich interaktiver, gruppenbasierter Prozesse zu systematisieren. Basierend auf dem entwickelten Modell wurde ein Softwareprototyp implementiert, der den gesamten Forschungsprozess begleiten kann. Dabei werden Gruppenformation, Feedback- und Reflexionsprozesse und das Peer Assessment mit Bildungstechnologien unterstützt. Die Entwicklungen wurden in einem qualitativen Experiment eingesetzt, um Systemwissen über die Möglichkeiten und Grenzen der digitalen Unterstützung von forschendem Lernen zu gewinnen.
Die 7. Fachtagung für Hochschuldidaktik, die 2016 erneut mit der DeLFI E-Learning Fachtagung Informatik stattfand, setzte das erfolgreiche Modell einer Tagung fort, die sich mit hochschuldidaktischen Fragen und der Gestaltung von Studiengängen der Informatik beschäftigt.
Thema der Tagung waren alle Fragen, die sich der Vermittlung von Informatikgegenständen im Hochschulbereich widmen. Dazu gehörten u.a.:
• fachdidaktische Konzepte der Vermittlung einzelner Informatikgegenstände
• methodische Lösungen, wie spezielle Lehr- und Lernformen, Durchführungskonzepte
• empirische Ergebnisse und Vergleichsstudien
• E-Learning-Ansätze, wenn sie ein erkennbares didaktisches Konzept verfolgen
• Studienkonzepte und Curricula, organisatorische Fragen, wie Gewinnung von Studierenden, Studieneingangsphase, Abbrecher.
Die Fachtagung widmete sich ausgewählten Fragestellungen dieses Themenkomplexes, die durch Vorträge ausgewiesener Experten, durch eingereichte Beiträge und durch Präsentationen und Poster intensiv behandelt wurden.
Unser besonderer Dank gilt dem Programmkomitee und den hier nicht genannten Helfern für ihren Einsatz bei der Vorbereitung und Durchführung der Tagung.
Die Fachtagungen HDI (Hochschuldidaktik Informatik) beschäftigen sich mit den unterschiedlichen Aspekten informatischer Bildung im Hochschulbereich. Neben den allgemeinen Themen wie verschiedenen Lehr- und Lernformen, dem Einsatz von Informatiksystemen in der Hochschullehre oder Fragen der Gewinnung von geeigneten Studierenden, deren Kompetenzerwerb oder auch der Betreuung der Studierenden widmet sich die HDI immer auch einem Schwerpunktthema.
Im Jahr 2021 war dies die Berücksichtigung von Diversität in der Lehre. Diskutiert wurden beispielsweise die Einbeziehung von besonderen fachlichen und überfachlichen Kompetenzen Studierender, der Unterstützung von Durchlässigkeit aus nichtakademischen Berufen, aber auch die Gestaltung inklusiver Lehr- und Lernszenarios, Aspekte des Lebenslangen Lernens oder sich an die Diversität von Studierenden adaptierte oder adaptierende Lehrsysteme.
Dieser Band enthält ausgewählte Beiträge der 9. Fachtagung 2021, die in besonderer Weise die Konferenz und die dort diskutierten Themen repräsentieren.
Services that operate over the Internet are under constant threat of being exposed to fraudulent use. Maintaining good user experience for legitimate users often requires the classification of entities as malicious or legitimate in order to initiate countermeasures. As an example, inbound email spam filters decide for spam or non-spam. They can base their decision on both the content of each email as well as on features that summarize prior emails received from the sending server. In general, discriminative classification methods learn to distinguish positive from negative entities. Each decision for a label may be based on features of the entity and related entities. When labels of related entities have strong interdependencies---as can be assumed e.g. for emails being delivered by the same user---classification decisions should not be made independently and dependencies should be modeled in the decision function. This thesis addresses the formulation of discriminative classification problems that are tailored for the specific demands of the following three Internet security applications. Theoretical and algorithmic solutions are devised to protect an email service against flooding of user inboxes, to mitigate abusive usage of outbound email servers, and to protect web servers against distributed denial of service attacks.
In the application of filtering an inbound email stream for unsolicited emails, utilizing features that go beyond each individual email's content can be valuable. Information about each sending mail server can be aggregated over time and may help in identifying unwanted emails. However, while this information will be available to the deployed email filter, some parts of the training data that are compiled by third party providers may not contain this information. The missing features have to be estimated at training time in order to learn a classification model. In this thesis an algorithm is derived that learns a decision function that integrates over a distribution of values for each missing entry. The distribution of missing values is a free parameter that is optimized to learn an optimal decision function.
The outbound stream of emails of an email service provider can be separated by the customer IDs that ask for delivery. All emails that are sent by the same ID in the same period of time are related, both in content and in label. Hijacked customer accounts may send batches of unsolicited emails to other email providers, which in turn might blacklist the sender's email servers after detection of incoming spam emails. The risk of being blocked from further delivery depends on the rate of outgoing unwanted emails and the duration of high spam sending rates. An optimization problem is developed that minimizes the expected cost for the email provider by learning a decision function that assigns a limit on the sending rate to customers based on the each customer's email stream.
Identifying attacking IPs during HTTP-level DDoS attacks allows to block those IPs from further accessing the web servers. DDoS attacks are usually carried out by infected clients that are members of the same botnet and show similar traffic patterns. HTTP-level attacks aim at exhausting one or more resources of the web server infrastructure, such as CPU time. If the joint set of attackers cannot increase resource usage close to the maximum capacity, no effect will be experienced by legitimate users of hosted web sites. However, if the additional load raises the computational burden towards the critical range, user experience will degrade until service may be unavailable altogether. As the loss of missing one attacker depends on block decisions for other attackers---if most other attackers are detected, not blocking one client will likely not be harmful---a structured output model has to be learned. In this thesis an algorithm is developed that learns a structured prediction decoder that searches the space of label assignments, guided by a policy.
Each model is evaluated on real-world data and is compared to reference methods. The results show that modeling each classification problem according to the specific demands of the task improves performance over solutions that do not consider the constraints inherent to an application.
Gerade in den letzten Jahren erfuhr Open Source Software (OSS) eine zunehmende Verbreitung und Popularität und hat sich in verschiedenen Anwendungsdomänen etabliert. Die Prozesse, welche sich im Kontext der OSS-Entwicklung (auch: OSSD – Open Source Software-Development) evolutionär herausgebildet haben, weisen in den verschiedenen OSS-Entwicklungsprojekten z.T. ähnliche Eigenschaften und Strukturen auf und auch die involvierten Entitäten, wie z.B. Artefakte, Rollen oder Software-Werkzeuge sind weitgehend miteinander vergleichbar. Dies motiviert den Gedanken, ein verallgemeinerbares Modell zu entwickeln, welches die generalisierbaren Entwicklungsprozesse im Kontext von OSS zu einem übertragbaren Modell abstrahiert. Auch in der Wissenschaftsdisziplin des Software Engineering (SE) wurde bereits erkannt, dass sich der OSSD-Ansatz in verschiedenen Aspekten erheblich von klassischen (proprietären) Modellen des SE unterscheidet und daher diese Methoden einer eigenen wissenschaftlichen Betrachtung bedürfen. In verschiedenen Publikationen wurden zwar bereits einzelne Aspekte der OSS-Entwicklung analysiert und Theorien über die zugrundeliegenden Entwicklungsmethoden formuliert, aber es existiert noch keine umfassende Beschreibung der typischen Prozesse der OSSD-Methodik, die auf einer empirischen Untersuchung existierender OSS-Entwicklungsprojekte basiert. Da dies eine Voraussetzung für die weitere wissenschaftliche Auseinandersetzung mit OSSD-Prozessen darstellt, wird im Rahmen dieser Arbeit auf der Basis vergleichender Fallstudien ein deskriptives Modell der OSSD-Prozesse hergeleitet und mit Modellierungselementen der UML formalisiert beschrieben. Das Modell generalisiert die identifizierten Prozesse, Prozessentitäten und Software-Infrastrukturen der untersuchten OSSD-Projekte. Es basiert auf einem eigens entwickelten Metamodell, welches die zu analysierenden Entitäten identifiziert und die Modellierungssichten und -elemente beschreibt, die zur UML-basierten Beschreibung der Entwicklungsprozesse verwendet werden. In einem weiteren Arbeitsschritt wird eine weiterführende Analyse des identifizierten Modells durchgeführt, um Implikationen, und Optimierungspotentiale aufzuzeigen. Diese umfassen beispielsweise die ungenügende Plan- und Terminierbarkeit von Prozessen oder die beobachtete Tendenz von OSSD-Akteuren, verschiedene Aktivitäten mit unterschiedlicher Intensität entsprechend der subjektiv wahrgenommenen Anreize auszuüben, was zur Vernachlässigung einiger Prozesse führt. Anschließend werden Optimierungszielstellungen dargestellt, die diese Unzulänglichkeiten adressieren, und ein Optimierungsansatz zur Verbesserung des OSSD-Modells wird beschrieben. Dieser Ansatz umfasst die Erweiterung der identifizierten Rollen, die Einführung neuer oder die Erweiterung bereits identifizierter Prozesse und die Modifikation oder Erweiterung der Artefakte des generalisierten OSS-Entwicklungsmodells. Die vorgestellten Modellerweiterungen dienen vor allem einer gesteigerten Qualitätssicherung und der Kompensation von vernachlässigten Prozessen, um sowohl die entwickelte Software- als auch die Prozessqualität im OSSD-Kontext zu verbessern. Desweiteren werden Softwarefunktionalitäten beschrieben, welche die identifizierte bestehende Software-Infrastruktur erweitern und eine gesamtheitlichere, softwaretechnische Unterstützung der OSSD-Prozesse ermöglichen sollen. Abschließend werden verschiedene Anwendungsszenarien der Methoden des OSS-Entwicklungsmodells, u.a. auch im kommerziellen SE, identifiziert und ein Implementierungsansatz basierend auf der OSS GENESIS vorgestellt, der zur Implementierung und Unterstützung des OSSD-Modells verwendet werden kann.
The goal of a Brain-Computer Interface (BCI) consists of the development of a unidirectional interface between a human and a computer to allow control of a device only via brain signals. While the BCI systems of almost all other groups require the user to be trained over several weeks or even months, the group of Prof. Dr. Klaus-Robert Müller in Berlin and Potsdam, which I belong to, was one of the first research groups in this field which used machine learning techniques on a large scale. The adaptivity of the processing system to the individual brain patterns of the subject confers huge advantages for the user. Thus BCI research is considered a hot topic in machine learning and computer science. It requires interdisciplinary cooperation between disparate fields such as neuroscience, since only by combining machine learning and signal processing techniques based on neurophysiological knowledge will the largest progress be made. In this work I particularly deal with my part of this project, which lies mainly in the area of computer science. I have considered the following three main points: <b>Establishing a performance measure based on information theory:</b> I have critically illuminated the assumptions of Shannon's information transfer rate for application in a BCI context. By establishing suitable coding strategies I was able to show that this theoretical measure approximates quite well to what is practically achieveable. <b>Transfer and development of suitable signal processing and machine learning techniques:</b> One substantial component of my work was to develop several machine learning and signal processing algorithms to improve the efficiency of a BCI. Based on the neurophysiological knowledge that several independent EEG features can be observed for some mental states, I have developed a method for combining different and maybe independent features which improved performance. In some cases the performance of the combination algorithm outperforms the best single performance by more than 50 %. Furthermore, I have theoretically and practically addressed via the development of suitable algorithms the question of the optimal number of classes which should be used for a BCI. It transpired that with BCI performances reported so far, three or four different mental states are optimal. For another extension I have combined ideas from signal processing with those of machine learning since a high gain can be achieved if the temporal filtering, i.e., the choice of frequency bands, is automatically adapted to each subject individually. <b>Implementation of the Berlin brain computer interface and realization of suitable experiments:</b> Finally a further substantial component of my work was to realize an online BCI system which includes the developed methods, but is also flexible enough to allow the simple realization of new algorithms and ideas. So far, bitrates of up to 40 bits per minute have been achieved with this system by absolutely untrained users which, compared to results of other groups, is highly successful.
In order to face the rapidly increasing need for computational resources of various scientific and engineering applications one has to think of new ways to make more efficient use of the worlds current computational resources. In this respect, the growing speed of wide area networks made a new kind of distributed computing possible: Metacomputing or (distributed) Grid computing. This is a rather new and uncharted field in computational science. The rapidly increasing speed of networks even outperforms the average increase of processor speed: Processor speeds double on average each 18 month whereas network bandwidths double every 9 months. Due to this development of local and wide area networks Grid computing will certainly play a key role in the future of parallel computing. This type of distributed computing, however, distinguishes from the traditional parallel computing in many ways since it has to deal with many problems not occurring in classical parallel computing. Those problems are for example heterogeneity, authentication and slow networks to mention only a few. Some of those problems, e.g. the allocation of distributed resources along with the providing of information about these resources to the application have been already attacked by the Globus software. Unfortunately, as far as we know, hardly any application or middle-ware software takes advantage of this information, since most parallelizing algorithms for finite differencing codes are implicitly designed for single supercomputer or cluster execution. We show that although it is possible to apply classical parallelizing algorithms in a Grid environment, in most cases the observed efficiency of the executed code is very poor. In this work we are closing this gap. In our thesis, we will - show that an execution of classical parallel codes in Grid environments is possible but very slow - analyze this situation of bad performance, nail down bottlenecks in communication, remove unnecessary overhead and other reasons for low performance - develop new and advanced algorithms for parallelisation that are aware of a Grid environment in order to generelize the traditional parallelization schemes - implement and test these new methods, replace and compare with the classical ones - introduce dynamic strategies that automatically adapt the running code to the nature of the underlying Grid environment. The higher the performance one can achieve for a single application by manual tuning for a Grid environment, the lower the chance that those changes are widely applicable to other programs. In our analysis as well as in our implementation we tried to keep the balance between high performance and generality. None of our changes directly affect code on the application level which makes our algorithms applicable to a whole class of real world applications. The implementation of our work is done within the Cactus framework using the Globus toolkit, since we think that these are the most reliable and advanced programming frameworks for supporting computations in Grid environments. On the other hand, however, we tried to be as general as possible, i.e. all methods and algorithms discussed in this thesis are independent of Cactus or Globus.
Flux-P
(2012)
Quantitative knowledge of intracellular fluxes in metabolic networks is invaluable for inferring metabolic system behavior and the design principles of biological systems. However, intracellular reaction rates can not often be calculated directly but have to be estimated; for instance, via 13C-based metabolic flux analysis, a model-based interpretation of stable carbon isotope patterns in intermediates of metabolism. Existing software such as FiatFlux, OpenFLUX or 13CFLUX supports experts in this complex analysis, but requires several steps that have to be carried out manually, hence restricting the use of this software for data interpretation to a rather small number of experiments. In this paper, we present Flux-P as an approach to automate and standardize 13C-based metabolic flux analysis, using the Bio-jETI workflow framework. Exemplarily based on the FiatFlux software, it demonstrates how services can be created that carry out the different analysis steps autonomously and how these can subsequently be assembled into software workflows that perform automated, high-throughput intracellular flux analysis of high quality and reproducibility. Besides significant acceleration and standardization of the data analysis, the agile workflow-based realization supports flexible changes of the analysis workflows on the user level, making it easy to perform custom analyses.
In a recent line of research, two familiar concepts from logic programming semantics (unfounded sets and splitting) were extrapolated to the case of epistemic logic programs. The property of epistemic splitting provides a natural and modular way to understand programs without epistemic cycles but, surprisingly, was only fulfilled by Gelfond's original semantics (G91), among the many proposals in the literature. On the other hand, G91 may suffer from a kind of self-supported, unfounded derivations when epistemic cycles come into play. Recently, the absence of these derivations was also formalised as a property of epistemic semantics called foundedness. Moreover, a first semantics proved to satisfy foundedness was also proposed, the so-called Founded Autoepistemic Equilibrium Logic (FAEEL). In this paper, we prove that FAEEL also satisfies the epistemic splitting property something that, together with foundedness, was not fulfilled by any other approach up to date. To prove this result, we provide an alternative characterisation of FAEEL as a combination of G91 with a simpler logic we called Founded Epistemic Equilibrium Logic (FEEL), which is somehow an extrapolation of the stable model semantics to the modal logic S5.
Modern biological analysis techniques supply scientists with various forms of data. One category of such data are the so called "expression data". These data indicate the quantities of biochemical compounds present in tissue samples. Recently, expression data can be generated at a high speed. This leads in turn to amounts of data no longer analysable by classical statistical techniques. Systems biology is the new field that focuses on the modelling of this information. At present, various methods are used for this purpose. One superordinate class of these methods is machine learning. Methods of this kind had, until recently, predominantly been used for classification and prediction tasks. This neglected a powerful secondary benefit: the ability to induce interpretable models. Obtaining such models from data has become a key issue within Systems biology. Numerous approaches have been proposed and intensively discussed. This thesis focuses on the examination and exploitation of one basic technique: decision trees. The concept of comparing sets of decision trees is developed. This method offers the possibility of identifying significant thresholds in continuous or discrete valued attributes through their corresponding set of decision trees. Finding significant thresholds in attributes is a means of identifying states in living organisms. Knowing about states is an invaluable clue to the understanding of dynamic processes in organisms. Applied to metabolite concentration data, the proposed method was able to identify states which were not found with conventional techniques for threshold extraction. A second approach exploits the structure of sets of decision trees for the discovery of combinatorial dependencies between attributes. Previous work on this issue has focused either on expensive computational methods or the interpretation of single decision trees a very limited exploitation of the data. This has led to incomplete or unstable results. That is why a new method is developed that uses sets of decision trees to overcome these limitations. Both the introduced methods are available as software tools. They can be applied consecutively or separately. That way they make up a package of analytical tools that usefully supplement existing methods. By means of these tools, the newly introduced methods were able to confirm existing knowledge and to suggest interesting and new relationships between metabolites.
Internetbasierte Informatiksysteme beeinflussen in steigendem Maße Situationen in unterschiedlichen Lebensbereichen. Kompetenzen zur Verwendung von Internetanwendungen und -diensten müssen explizit erworben werden, weil damit ein notwendiger Einblick in nicht beobachtbare Abläufe und nicht offen sichtbare Strukturen verbunden ist. Bisher gibt es Vorschläge für die Gestaltung schulischer Lehr-Lernprozesse zu ausgewählten Teilaspekten des Internets. Es fehlt eine systematische Analyse des Bildungsbedarfs und ein daraus resultierendes Unterrichtsmodell. In dieser Arbeit wird ein Gesamtkonzept für den Informatikunterricht in der Sekundarstufe II vorgestellt, das zu zielgerichteter und verantwortungsvoller Anwendung des Internets beiträgt. Die vorliegende Arbeit umfasst den Prozess von der Analyse erforderlicher Kompetenzen bis zur Realisierung von Lehr-Lernprozessen im Informatikunterricht in der Sekundarstufe II. Es werden der Beitrag der Informatik zu identifizierten Kompetenzen untersucht und Bildungsanforderungen bestimmt. Bildungsempfehlungen und Forschungsergebnisse zu erfolgreichen Unterrichtseinheiten werden im Hinblick auf die Bildungsziele analysiert. Der Informatikunterricht unterstützt die Kompetenzentwicklung zu internetbasierten digitalen Medien. Es wird die Entwicklung eines Unterrichtsmodells zu Internetworking beschrieben. Dazu wird der Ansatz der Didaktischen Systeme untersucht, weiter entwickelt und auf den Bereich Internetworking übertragen. Der theoretische Ansatz wird dazu in vier Unterrichtsprojekten zu Internetworking in der Praxis realisiert. Beziehungen zwischen Fachkonzepten zu Internetworking werden untersucht und durch Wissensstrukturen zur Planung von Unterrichtsprojekten eingesetzt und in der Praxis erprobt. Die Beschreibung von Lernaktivitäten erfolgt auf der Basis von Aufgabenklassen, die das notwendige Wissen zur Bearbeitung einer Aufgabenstellung repräsentieren. Auf der Grundlage des Ablaufs der Aufgabenbearbeitung werden Eigenschaften von Aufgaben beschrieben und zu deren Gestaltung nutzbar gemacht. Bisher nicht durchführbare Tätigkeiten im Unterricht werden durch die Entwicklung der Lernsoftware Filius ermöglicht. Die Reduktion der komplexen Wirklichkeit durch Simulation realer internetbasierter Informatiksysteme und die Auswahl geeigneter Sichten auf den Untersuchungsgegenstand werden mit Ergebnissen der Informatikdidaktik begründet. Unterrichtsprojekte zu den Zielen werden durchgeführt, um Lehr-Lernprozesse zu erkunden und das entwickelte Didaktische System zu erproben. Ausgehend von der theoretischen Fundierung erfolgt die praktische Realisierung von Lehr-Lernprozessen. Zur Erprobung im Informatikunterricht der Sekundarstufe II in Nordrhein-Westfalen werden Minimalziele aufgrund der Lehrvorgaben bestimmt. Die methodische Gestaltung in der Erprobung erfolgt unter Berücksichtigung der Vorgaben für den Informatikunterricht und allgemeinen Anforderungen der Fachdidaktik. Handlungsorientierte Unterrichtsmittel werden ausgewählt und in der Praxis zur Untersuchung der Lehr-Lernprozesse verwendet. Im Unterricht identifizierte Lernschwierigkeiten führen zur Modifikation der Wissensstrukturen und werden im Entwicklungsprozess von Filius berücksichtigt. Die Erkenntnisse aus Unterrichtsprojekten werden genutzt, um zu bestimmen, zu welchen Aufgabenklassen weitere Aufgaben erforderlich sind und inwieweit das aus den identifizierten Merkmalen abgeleitete Vorgehen zur Entwicklung niveaubestimmender Aufgaben genutzt werden kann. Die Erprobungen bestätigen die Tragfähigkeit des Didaktischen Systems Internetworking und leisten mit der Implementierung in der Praxis einen Beitrag zur Untersuchung von Kompetenzentwicklung im Informatikunterricht. Mit dem Didaktischen System Internetworking wird ein theoretisch fundiertes und empirisch erprobtes Unterrichtsmodell zur Entwicklung von Kompetenzen zur Einrichtung und Anwendung internetbasierter Informatiksysteme beschrieben.
Die Tagung HDI 2014 in Freiburg zur Hochschuldidaktik der Informatik HDI wurde erneut vom Fachbereich Informatik und Ausbildung / Didaktik der Informatik (IAD) in der Gesellschaft für Informatik e. V. (GI) organisiert. Sie dient den Lehrenden der Informatik in Studiengängen an Hochschulen als Forum der Information und des Austauschs über neue didaktische Ansätze und bildungspolitische Themen im Bereich der Hochschulausbildung aus der fachlichen Perspektive der Informatik.
Die HDI 2014 ist nun bereits die sechste Ausgabe der HDI. Für sie wurde das spezielle Motto „Gestalten und Meistern von Übergängen“ gewählt. Damit soll ein besonderes Augenmerk auf die Übergänge von Schule zum Studium, vom Bachelor zum Master, vom Studium zur Promotion oder vom Studium zur Arbeitswelt gelegt werden.
Answer Set Programming (ASP) is an emerging paradigm for declarative programming, in which a computational problem is specified by a logic program such that particular models, called answer sets, match solutions. ASP faces a growing range of applications, demanding for high-performance tools able to solve complex problems. ASP integrates ideas from a variety of neighboring fields. In particular, automated techniques to search for answer sets are inspired by Boolean Satisfiability (SAT) solving approaches. While the latter have firm proof-theoretic foundations, ASP lacks formal frameworks for characterizing and comparing solving methods. Furthermore, sophisticated search patterns of modern SAT solvers, successfully applied in areas like, e.g., model checking and verification, are not yet established in ASP solving. We address these deficiencies by, for one, providing proof-theoretic frameworks that allow for characterizing, comparing, and analyzing approaches to answer set computation. For another, we devise modern ASP solving algorithms that integrate and extend state-of-the-art techniques for Boolean constraint solving. We thus contribute to the understanding of existing ASP solving approaches and their interconnections as well as to their enhancement by incorporating sophisticated search patterns. The central idea of our approach is to identify atomic as well as composite constituents of a propositional logic program with Boolean variables. This enables us to describe fundamental inference steps, and to selectively combine them in proof-theoretic characterizations of various ASP solving methods. In particular, we show that different concepts of case analyses applied by existing ASP solvers implicate mutual exponential separations regarding their best-case complexities. We also develop a generic proof-theoretic framework amenable to language extensions, and we point out that exponential separations can likewise be obtained due to case analyses on them. We further exploit fundamental inference steps to derive Boolean constraints characterizing answer sets. They enable the conception of ASP solving algorithms including search patterns of modern SAT solvers, while also allowing for direct technology transfers between the areas of ASP and SAT solving. Beyond the search for one answer set of a logic program, we address the enumeration of answer sets and their projections to a subvocabulary, respectively. The algorithms we develop enable repetition-free enumeration in polynomial space without being intrusive, i.e., they do not necessitate any modifications of computations before an answer set is found. Our approach to ASP solving is implemented in clasp, a state-of-the-art Boolean constraint solver that has successfully participated in recent solver competitions. Although we do here not address the implementation techniques of clasp or all of its features, we present the principles of its success in the context of ASP solving.
Quantified Boolean formulas (QBFs) play an important role in theoretical computer science. QBF extends propositional logic in such a way that many advanced forms of reasoning can be easily formulated and evaluated. In this dissertation we present our ZQSAT, which is an algorithm for evaluating quantified Boolean formulas. ZQSAT is based on ZBDD: Zero-Suppressed Binary Decision Diagram , which is a variant of BDD, and an adopted version of the DPLL algorithm. It has been implemented in C using the CUDD: Colorado University Decision Diagram package. The capability of ZBDDs in storing sets of subsets efficiently enabled us to store the clauses of a QBF very compactly and let us to embed the notion of memoization to the DPLL algorithm. These points led us to implement the search algorithm in such a way that we could store and reuse the results of all previously solved subformulas with a little overheads. ZQSAT can solve some sets of standard QBF benchmark problems (known to be hard for DPLL based algorithms) faster than the best existing solvers. In addition to prenex-CNF, ZQSAT accepts prenex-NNF formulas. We show and prove how this capability can be exponentially beneficial.