004 Datenverarbeitung; Informatik
Refine
Has Fulltext
- yes (207) (remove)
Year of publication
Document Type
- Article (110)
- Doctoral Thesis (64)
- Conference Proceeding (13)
- Master's Thesis (8)
- Postprint (6)
- Monograph/Edited Volume (2)
- Bachelor Thesis (1)
- Habilitation Thesis (1)
- Moving Images (1)
- Preprint (1)
Keywords
- Informatik (17)
- Didaktik (15)
- Hochschuldidaktik (14)
- Ausbildung (13)
- Maschinelles Lernen (7)
- Antwortmengenprogrammierung (5)
- Computer Science Education (5)
- Machine Learning (5)
- answer set programming (5)
- E-Learning (4)
- education (4)
- higher education (4)
- Answer Set Programming (3)
- Competence Measurement (3)
- DPLL (3)
- Informatikdidaktik (3)
- Secondary Education (3)
- didactics (3)
- informatics (3)
- Algorithmen (2)
- Algorithms (2)
- Assessment (2)
- Automatisches Beweisen (2)
- Big Data (2)
- Bildverarbeitung (2)
- Code (2)
- Competence Modelling (2)
- Computational thinking (2)
- Computer Science (2)
- Computergrafik (2)
- Computersicherheit (2)
- Constraint Solving (2)
- Data Privacy (2)
- Deduction (2)
- EEG (2)
- FMC (2)
- HCI (2)
- ICA (2)
- Informatics (2)
- Informatics Education (2)
- Informatics Modelling (2)
- Informatics System Application (2)
- Informatics System Comprehension (2)
- Informatikstudium (2)
- Key Competencies (2)
- Klausellernen (2)
- Knowledge Representation and Reasoning (2)
- Künstliche Intelligenz (2)
- Logic Programming (2)
- Logics (2)
- Mensch-Computer-Interaktion (2)
- Modell (2)
- Modellierung (2)
- Ontologie (2)
- Optimization (2)
- Planing (2)
- Prozessmodellierung (2)
- Relevanz (2)
- SAT (2)
- Semantic Web (2)
- Systemstruktur (2)
- Texturen (2)
- Theorembeweisen (2)
- Unifikation (2)
- Visualisierung (2)
- Vorhersage (2)
- abstraction (2)
- code (2)
- complexity (2)
- computational thinking (2)
- computer graphics (2)
- computer science education (2)
- human computer interaction (2)
- image processing (2)
- machine learning (2)
- model (2)
- non-photorealistic rendering (2)
- relevance (2)
- systems biology (2)
- textures (2)
- theorem (2)
- virtual 3D city models (2)
- virtuelle 3D-Stadtmodelle (2)
- visualization (2)
- 'Peer To Peer' (1)
- 21st century skills, (1)
- 3D Computer Grafik (1)
- 3D Computer Graphics (1)
- 3D Drucken (1)
- 3D Linsen (1)
- 3D Semiotik (1)
- 3D Visualisierung (1)
- 3D computer graphics (1)
- 3D lenses (1)
- 3D printing (1)
- 3D semiotics (1)
- 3D visualization (1)
- 3D-Stadtmodelle (1)
- 3d city models (1)
- ABRACADABRA (1)
- ADFS (1)
- ARCS Modell (1)
- Abbrecherquote (1)
- Abstraktion (1)
- Accepting Grammars (1)
- Achievement (1)
- Ackerschmalwand (1)
- Active Directory Federation Services (1)
- Active Evaluation (1)
- Activity Theory (1)
- Activity-orientated Learning (1)
- Adversarial Learning (1)
- Aktive Evaluierung (1)
- Akzeptierende Grammatiken (1)
- Algorithmenablaufplanung (1)
- Algorithmenkonfiguration (1)
- Algorithmenselektion (1)
- Alignment (1)
- Angewandte Spieltheorie (1)
- Anisotroper Kuwahara Filter (1)
- Anleitung (1)
- Antwortmengen Programmierung (1)
- Applied Game Theory (1)
- Arabidopsis thaliana (1)
- Arduino (1)
- Artificial Intelligence (1)
- Aspect-Oriented Programming (1)
- Aspektorientierte Programmierung (1)
- Asynchrone Schaltung (1)
- Audience Response Systeme (1)
- Aufzählung (1)
- Ausreissererkennung (1)
- Austria (1)
- Authentifizierung (1)
- Authorization (1)
- Autismus (1)
- Automated Theorem Proving (1)
- Autorisierung (1)
- BCH (1)
- BCI (1)
- BSS (1)
- Bachelorstudierende der Informatik (1)
- Baumweite (1)
- Behavior (1)
- Benutzeroberfläche (1)
- Berührungseingaben (1)
- Betrachtungsebenen (1)
- Beweistheorie (1)
- Bilddatenanalyse (1)
- Bildung (1)
- Binäres Entscheidungsdiagramm (1)
- Biocomputing (1)
- Bioelektrisches Signal (1)
- Bioinformatik (1)
- Blended Learning (1)
- Bloom’s Taxonomy (1)
- Boolean constraint solver (1)
- Brain Computer Interface (1)
- Business Process Models (1)
- CS concepts (1)
- CSC (1)
- Cactus (1)
- Capability approach (1)
- Challenges (1)
- Choreographien (1)
- CityGML (1)
- Classification (1)
- Clause Learning (1)
- Cloud Computing (1)
- Cloud computing (1)
- Clusteranalyse (1)
- Codierung (1)
- Cognitive Skills (1)
- Common Spatial Pattern (1)
- Competences (1)
- Competencies (1)
- Complementary Circuits (1)
- Compliance (1)
- Composition (1)
- Computational Complexity (1)
- Computational Thinking (1)
- Computer Science in Context (1)
- Computergestützes Training (1)
- Computing (1)
- Conceptual (1)
- Constraint (1)
- Constraint-Programmierung (1)
- Constructive solid geometry (1)
- Contest (1)
- Contextualisation (1)
- Contradictions (1)
- Controlled Derivations (1)
- Covariate Shift (1)
- Curriculum (1)
- Curriculum Development (1)
- DDoS (1)
- DNA (1)
- DNA computing (1)
- DNS (1)
- Data Analysis (1)
- Data Management (1)
- Databases (1)
- Datenschutz (1)
- Deep Learning (1)
- Defining characteristics of physical computing (1)
- Dempster-Shafer-Theorie (1)
- Dempster–Shafer theory (1)
- Description Logics (1)
- Deskriptive Logik (1)
- Diagonalisierung (1)
- Didaktik der Informatik (1)
- Didaktische Konzepte (1)
- Dienstkomposition (1)
- Dienstplattform (1)
- Differenz von Gauss Filtern (1)
- Digital Competence (1)
- Digital Education (1)
- Digital Game Based Learning (1)
- Digital Revolution (1)
- Distributed Computing (1)
- Diversität (1)
- Dreidimensionale Computergraphik (1)
- Dynamic Programming (1)
- Dynamic assessment (1)
- Dynamische Programmierung (1)
- Dynamische Rekonfiguration (1)
- Early Literacy (1)
- Echtzeitanwendung (1)
- Educational Standards (1)
- Educational software (1)
- Eingabegenauigkeit (1)
- Eisenbahnnetz (1)
- Elektroencephalographie (1)
- Embedded Systems (1)
- Emotionen (1)
- Emotionsforschung (1)
- Enterprise Search (1)
- Entscheidungsbäume (1)
- Entwurf (1)
- Entwurfsmuster (1)
- Entwurfsmuster für SOA-Sicherheit (1)
- Entwurfsraumexploration (1)
- Equilibrium logic (1)
- Erfüllbarkeit einer Formel der Aussagenlogik (1)
- Erfüllbarkeitsproblem (1)
- Error Estimation (1)
- Error-Detection Circuits (1)
- Euclid’s algorithm (1)
- European Bioinformatics Institute (1)
- Evidenztheorie (1)
- Exploration (1)
- Exponential Time Hypothesis (1)
- Exponentialzeit Hypothese (1)
- FMC-QE (1)
- FOSS (1)
- FPGA (1)
- Facebook (1)
- Feature Combination (1)
- Feedback (1)
- Fehlende Daten (1)
- Fehlererkennung (1)
- Fehlerkorrektur (1)
- Fehlerschätzung (1)
- Fehlvorstellung (1)
- Fibonacci numbers (1)
- Flussgesteuerter Bilateraler Filter (1)
- Focus+Context Visualization (1)
- Fokus-&-Kontext Visualisierung (1)
- Formeln der quantifizierten Aussagenlogik (1)
- Function (1)
- Fundamental Ideas (1)
- GPU (1)
- Game-based learning (1)
- Gebäudemodelle (1)
- Gehirn-Computer-Schnittstelle (1)
- Geländemodelle (1)
- Generalisierung (1)
- Geodaten (1)
- Geometrieerzeugung (1)
- Geovisualisierung (1)
- Geschäftsprozessmodelle (1)
- Gesichtsausdruck (1)
- Gesteuerte Ableitungen (1)
- Gleichheit (1)
- Globus (1)
- Grammar Systems (1)
- Grammatikalische Inferenz (1)
- Grammatiksysteme (1)
- Graph-basiertes Ranking (1)
- Graphensuche (1)
- Graphfärbung (1)
- Grid (1)
- Grid Computing (1)
- HDI (1)
- Hardware-Software-Co-Design (1)
- Hauptkomponentenanalyse (1)
- Heterogenität (1)
- High-Level Synthesis (1)
- Hochschullehre (1)
- Hochschulsystem (1)
- I/O-effiziente Algorithmen (1)
- ICT (1)
- ICT Competence (1)
- ICT competencies (1)
- ICT skills (1)
- IT security (1)
- IT-Security (1)
- IT-Sicherheit (1)
- Informatik-Studiengänge (1)
- Informatiksystem (1)
- Informatikunterricht (1)
- Informatikvoraussetzungen (1)
- Information Transfer Rate (1)
- Informationsextraktion (1)
- Infrastruktur (1)
- Inkonsistenz (1)
- Inquiry-based Learning (1)
- Integration (1)
- Interactive Rendering (1)
- Interactive system (1)
- Interaktionsmodel (1)
- Interaktionsmodellierung (1)
- Interaktionstechniken (1)
- Interaktives Rendering (1)
- Interaktives System (1)
- Interface design (1)
- Internet Security (1)
- Internet applications (1)
- Internet-Sicherheit (1)
- Internetanwendungen (1)
- Intuition (1)
- Java Security Framework (1)
- Kartografisches Design (1)
- Kern-PCA (1)
- Kernmethoden (1)
- Klassifikation (1)
- Klassifikator-Kalibrierung (1)
- Kommunikation (1)
- Kompetenz (1)
- Kompetenzerwerb (1)
- Komplexität (1)
- Komplexitätsbewältigung (1)
- Komplexitätstheorie (1)
- Komposition (1)
- Konnektionskalkül (1)
- Konzeptionell (1)
- Kryptographie (1)
- Kybernetik (1)
- LMS (1)
- Landmarken (1)
- Laser Cutten (1)
- Learners (1)
- Learning Fields (1)
- Learning ecology (1)
- Learning interfaces development (1)
- Learning with ICT (1)
- Lebenslanges Lernen (1)
- Leftmost Derivations (1)
- Lehrer (1)
- Lehrer*innenbildung (1)
- Leistungsfähigkeit (1)
- Leistungsvorhersage (1)
- Lern-App (1)
- Lernmotivation (1)
- Lernsoftware (1)
- Liguistisch (1)
- Linksableitungen (1)
- Logarithm (1)
- Logikkalkül (1)
- Logiksynthese (1)
- Lower Bounds (1)
- Lower Secondary Level (1)
- MEG (1)
- MOOCs (1)
- Magnetoencephalographie (1)
- Malware (1)
- Massive Open Online Courses (1)
- Matrizen-Eigenwertaufgabe (1)
- Measurement (1)
- Megamodel (1)
- Megamodell (1)
- Mehrklassen-Klassifikation (1)
- Metamodell (1)
- Middleware (1)
- Migration (1)
- Mischmodelle (1)
- Mischung <Signalverarbeitung> (1)
- Mobiles Lernen (1)
- Mobilgeräte (1)
- Model Based Engineering (1)
- Model Checking (1)
- Model Driven Architecture (1)
- Model Management (1)
- Model-Driven Engineering (1)
- Modeling (1)
- Modell Management (1)
- Modell-driven Security (1)
- Modell-getriebene Sicherheit (1)
- Modellbasiert (1)
- Modellgetriebene Architektur (1)
- Modellgetriebene Entwicklung (1)
- Molekulare Bioinformatik (1)
- Multi Task Learning (1)
- Multi-Class (1)
- Multi-Task-Lernen (1)
- Multiprocessor (1)
- Multiprozessor (1)
- Music Technology (1)
- NUI (1)
- Natural Science Education (1)
- Navigation (1)
- Netzwerk (1)
- Netzwerke (1)
- Neuronales Netz (1)
- New On-Line Error-Detection Methode (1)
- Next Generation Network (1)
- Nicht-photorealistisches Rendering (1)
- Nichtfotorealistische Bildsynthese (1)
- NoSQL (1)
- Norway (1)
- Novice programmers (1)
- Nutzungsinteresse (1)
- Objektive Schwierigkeit (1)
- Omega (1)
- Onlinelehre (1)
- Ontologies (1)
- Ontology (1)
- Open Source (1)
- OpenOLAT (1)
- Optimierungsproblem (1)
- Owner-Retained Access Control (ORAC) (1)
- Parallel Programming (1)
- Parallele Datenverarbeitung (1)
- Paralleles Rechnen (1)
- Parallelrechner (1)
- Parameterized Complexity (1)
- Parametrisierte Komplexität (1)
- Parsing (1)
- Patterns (1)
- Pedagogical content knowledge (1)
- Peer-to-Peer-Netz ; GRID computing ; Zuverlässigkeit ; Web Services ; Betriebsmittelverwaltung ; Migration (1)
- Performance (1)
- Performance Prediction (1)
- Physical Science (1)
- Platzierung (1)
- Policy Enforcement (1)
- Policy Languages (1)
- Policy Sprachen (1)
- Pre-RS Traceability (1)
- Prediction Game (1)
- Predictive Models (1)
- Preprocessing (1)
- Problem Solving (1)
- Probleme in der Studie (1)
- Problemlösen (1)
- Process (1)
- Process modeling (1)
- Professoren (1)
- Programmierung (1)
- Proof Theory (1)
- Prozess (1)
- Prozess Verbesserung (1)
- Prozesse (1)
- Prozessmodell (1)
- Prädiktionsspiel (1)
- Präferenzen (1)
- Quantenkryptographie (1)
- Quantified Boolean Formula (QBF) (1)
- Quantitative Modeling (1)
- Quantitative Modellierung (1)
- Queuing Theory (1)
- Recommendations for CS-Curricula in Higher Education (1)
- Reconfigurable (1)
- Rekonfiguration (1)
- Rendering (1)
- Reparatur (1)
- SOA Security Pattern (1)
- SSO (1)
- STG decomposition (1)
- STG-Dekomposition (1)
- Sample Selection Bias (1)
- Satisfiability (1)
- Scalability (1)
- Scene graph systems (1)
- Schlüsselkompetenzen (1)
- Security Modelling (1)
- Segmentierung (1)
- Selektion (1)
- Selektionsbias (1)
- Self-Checking Circuits (1)
- Semantic Search (1)
- Semantik Web (1)
- Semantische Suche (1)
- Sensors (1)
- Service Creation (1)
- Service Delivery Platform (1)
- Service convergence (1)
- Service-Orientierte Architekturen (1)
- Service-oriented Architectures (1)
- Shader (1)
- Sicherheitsmodellierung (1)
- Signal Processing (1)
- Signalquellentrennung (1)
- Signaltrennung (1)
- Simultane Diagonalisierung (1)
- Single Sign On (1)
- Single Trial Analysis (1)
- Skalierbarkeit (1)
- Skelettberechnung (1)
- Small Private Online Courses (1)
- Social (1)
- Software (1)
- Software Engineering (1)
- Software architecture (1)
- Softwarearchitektur (1)
- Softwareentwicklung (1)
- Spam (1)
- Spam Filtering (1)
- Spam-Erkennung (1)
- Spam-Filter (1)
- Spam-Filtering (1)
- Spatio-Spectral Filter (1)
- Spawning (1)
- Sprachdesign (1)
- Statistical Tests (1)
- Statistikprogramm R (1)
- Statistische Tests (1)
- Stilisierung (1)
- Strategie (1)
- Structuring (1)
- Strukturierung (1)
- Studentenerwartungen (1)
- Studentenhaltungen (1)
- Studentenjobs (1)
- Studienabbrecher (1)
- Studiendauer (1)
- Suche (1)
- Synthese (1)
- System Biologie (1)
- System structure (1)
- Systembiologie (1)
- Systementwurf (1)
- Szenengraph (1)
- TPTP (1)
- Tasks (1)
- Teacher perceptions (1)
- Teachers (1)
- Teaching information security (1)
- Technology proficiency (1)
- Telekommunikation (1)
- Temporal Logic (1)
- Temporallogik (1)
- Temporäre Anbindung (1)
- Terminologische Logik (1)
- Terminology (1)
- Test (1)
- Tests (1)
- Theoretische Informatik (1)
- Theoretischen Vorlesungen (1)
- Theory (1)
- Time Augmented Petri Nets (1)
- Tool (1)
- Traceability (1)
- Tracking (1)
- Transformation (1)
- Treewidth (1)
- Unabhängige Komponentenanalyse (1)
- Universität Bagdad (1)
- Universität Potsdam (1)
- Universitätseinstellungen (1)
- Untere Schranken (1)
- Unvollständigkeit (1)
- Usage Interest (1)
- User Experience (1)
- VM (1)
- Verhalten (1)
- Verifikation (1)
- Verletzung Auflösung (1)
- Verletzung Erklärung (1)
- Verteiltes Rechnen (1)
- Verteilungsunterschied (1)
- Violation Explanation (1)
- Violation Resolution (1)
- Virtual Reality (1)
- Virtuelles 3D Stadtmodell (1)
- Vocational Education (1)
- Vorhersagemodelle (1)
- Wahrnehmung (1)
- Wahrnehmung von Arousal (1)
- Wahrnehmungsunterschiede (1)
- Warteschlangentheorie (1)
- Web Services (1)
- Web Sites (1)
- Web of Data (1)
- Webseite (1)
- Weiterbildung (1)
- Well-structuredness (1)
- Wirtschaftsinformatik (1)
- Wissensrepräsentation und -verarbeitung (1)
- Wissensrepräsentation und Schlussfolgerung (1)
- Wohlstrukturiertheit (1)
- Workflow (1)
- Young People (1)
- ZQSA (1)
- ZQSAT (1)
- Zeitbehaftete Petri Netze (1)
- Zero-Suppressed Binary Decision Diagram (ZDD) (1)
- Zuverlässigkeitsanalyse (1)
- adaptiv (1)
- adaptive (1)
- algorithm configuration (1)
- algorithm scheduling (1)
- algorithm selection (1)
- analogical thinking (1)
- animated PCA (1)
- animierte PCA (1)
- anisotropic Kuwahara filter (1)
- approximate joint diagonalization (1)
- argument mining (1)
- argumentation structure (1)
- arithmethische Prozeduren (1)
- arithmetic procedures (1)
- arousal perception (1)
- artificial intelligence (1)
- assistive Technologien (1)
- assistive technologies (1)
- asynchronous circuit (1)
- authentication (1)
- automated theorem proving (1)
- automatic theorem prover (1)
- automatisierter Theorembeweiser (1)
- bild (1)
- binary representation (1)
- binary search (1)
- bio-computing (1)
- blind source separation (1)
- building models (1)
- business informatics (1)
- cartographic design (1)
- changing the study field (1)
- changing the university (1)
- choreographies (1)
- classifier calibration (1)
- classroom language (1)
- clause learning (1)
- clustering (1)
- cognitive modifiability (1)
- coherence-enhancing filtering (1)
- communication (1)
- competence (1)
- competencies (1)
- competency (1)
- comprehension (1)
- computational biology (1)
- computational methods (1)
- computer science teachers (1)
- computer security (1)
- computergestützte Methoden (1)
- connection calculus (1)
- constraint (1)
- constraint programming (1)
- cryptography (1)
- cs4fn (1)
- curriculum theory (1)
- decision trees (1)
- design (1)
- design space exploration (1)
- didaktisches Konzept (1)
- difference of Gaussians (1)
- digitale Bildung (1)
- digitale Hochschullehre (1)
- digitally-enabled pedagogies (1)
- divide and conquer (1)
- dropout (1)
- dynamic (1)
- dynamic classification (1)
- dynamic reconfiguration (1)
- dynamisch (1)
- dynamische Klassifikation (1)
- e-Learning (1)
- e-mentoring (1)
- education and public policy (1)
- educational programming (1)
- educational systems (1)
- edutainment (1)
- email spam detection (1)
- emotion (1)
- emotion representation (1)
- emotion research (1)
- enterprise search (1)
- entity alignment (1)
- enumeration (1)
- environments (1)
- epistemic logic programs (1)
- epistemic specifications (1)
- equality (1)
- error correction (1)
- error detection (1)
- evidence theory (1)
- explicit negation (1)
- exponentiation (1)
- external memory algorithms (1)
- face tracking (1)
- facial expression (1)
- flow-based bilateral filter (1)
- freie Daten (1)
- freie Software (1)
- fun (1)
- generalization (1)
- geometry generation (1)
- geospatial data (1)
- geovisualization (1)
- global constraints (1)
- globale Constraints (1)
- grammar inference (1)
- graph clustering (1)
- graph-based ranking (1)
- graph-search (1)
- hardware-software-codesign (1)
- high school (1)
- higher (1)
- image (1)
- image data analysis (1)
- incompleteness (1)
- inconsistency (1)
- independent component analysis (1)
- informal and formal learning (1)
- informatics education (1)
- information extraction (1)
- infrastructure (1)
- innovation (1)
- input accuracy (1)
- interaction modeling (1)
- interaction techniques (1)
- interactive course (1)
- interactive workshop (1)
- intuition (1)
- kernel PCA (1)
- kernel methods (1)
- key competences in physical computing (1)
- key competencies (1)
- kinaesthetic teaching (1)
- klinisch-praktischer Unterricht (1)
- konvergente Dienste (1)
- landmarks (1)
- language design (1)
- leanCoP (1)
- learning (1)
- linear code (1)
- linearer Code (1)
- linguistic (1)
- logic programming (1)
- logic synthesis (1)
- logical calculus (1)
- logical signaling networks (1)
- logische Ergänzung (1)
- logische Programmierung (1)
- logische Signalnetzwerke (1)
- malware detection (1)
- map/reduce (1)
- maschinelles Lernen (1)
- maschninelles Lernen (1)
- mediated learning experience (1)
- medical (1)
- medizinisch (1)
- meta model (1)
- misconception (1)
- mixture models (1)
- mobile devices (1)
- mobile learning (1)
- mobile technologies and apps (1)
- mobiles Lernen (1)
- model-based (1)
- model-driven architecture (1)
- molecular networks (1)
- molekulare Netzwerke (1)
- multi core data processing (1)
- multi-class classification (1)
- multiuser (1)
- navigation (1)
- network (1)
- networks (1)
- networks-on-chip (1)
- neue Online-Fehlererkennungsmethode (1)
- nichtlineare ICA (1)
- nichtlineare PCA (NLPCA) (1)
- nichtlineare Projektionen (1)
- non-monotonic reasoning (1)
- nonlinear ICA (1)
- nonlinear PCA (NLPCA) (1)
- nonlinear projections (1)
- objective difficulty (1)
- omega (1)
- on-chip (1)
- online assistance (1)
- open learning (1)
- open source (1)
- operating system (1)
- organisational evolution (1)
- outlier detection (1)
- overcomplete ICA (1)
- paper prototyping (1)
- parallel solving (1)
- paralleles Lösen (1)
- parameter (1)
- parsing (1)
- pattern recognition (1)
- pedagogy (1)
- perception (1)
- perception differences (1)
- personal (1)
- personal response systems (1)
- philosophical foundation of informatics pedagogy (1)
- physical computing tools (1)
- placement (1)
- policy evaluation (1)
- pre-primary level (1)
- prediction (1)
- preferences (1)
- preprocessing (1)
- primary education (1)
- primary level (1)
- priorities (1)
- problem-solving (1)
- process (1)
- process improvement (1)
- process model (1)
- process modelling (1)
- professional development (1)
- professors (1)
- programming (1)
- programming in context (1)
- proving (1)
- quantum cryptography (1)
- railway network (1)
- real-time application (1)
- reconfiguration (1)
- rekonfigurierbar (1)
- reliability assessment (1)
- repair (1)
- robust ICA (1)
- robuste ICA (1)
- scheduling (1)
- search (1)
- secondary computer science education (1)
- secondary education (1)
- segmentation (1)
- selbstprüfende Schaltungen (1)
- selection (1)
- self-efficacy (1)
- semantisches Netz (1)
- shader (1)
- sign language (1)
- skeletonization (1)
- social media (1)
- software (1)
- software development (1)
- software engineering (1)
- speed independence (1)
- statistics program R (1)
- structured output prediction (1)
- strukturierte Vorhersage (1)
- student activation (1)
- student experience (1)
- student perceptions (1)
- students’ conceptions (1)
- students’ knowledge (1)
- study problems (1)
- stylization (1)
- teacher competencies (1)
- teacher training (1)
- teachers (1)
- teaching informatics in general education (1)
- technische Rahmenbedingungen (1)
- temporary binding (1)
- terrain models (1)
- test (1)
- tools (1)
- topics (1)
- touch input (1)
- tptp (1)
- tracing (1)
- transformation (1)
- tutorial section (1)
- unification (1)
- user interfaces (1)
- user-centred (1)
- verification (1)
- virtual 3D city model (1)
- virtual machine (1)
- virtual mobility (1)
- virtual reality (1)
- überbestimmte ICA (1)
- ‘unplugged’ computing (1)
Institute
- Institut für Informatik und Computational Science (207) (remove)
In the early days of computer graphics, research was mainly driven by the goal to create realistic synthetic imagery. By contrast, non-photorealistic computer graphics, established as its own branch of computer graphics in the early 1990s, is mainly motivated by concepts and principles found in traditional art forms, such as painting, illustration, and graphic design, and it investigates concepts and techniques that abstract from reality using expressive, stylized, or illustrative rendering techniques. This thesis focuses on the artistic stylization of two-dimensional content and presents several novel automatic techniques for the creation of simplified stylistic illustrations from color images, video, and 3D renderings. Primary innovation of these novel techniques is that they utilize the smooth structure tensor as a simple and efficient way to obtain information about the local structure of an image. More specifically, this thesis contributes to knowledge in this field in the following ways. First, a comprehensive review of the structure tensor is provided. In particular, different methods for integrating the minor eigenvector field of the smoothed structure tensor are developed, and the superiority of the smoothed structure tensor over the popular edge tangent flow is demonstrated. Second, separable implementations of the popular bilateral and difference of Gaussians filters that adapt to the local structure are presented. These filters avoid artifacts while being computationally highly efficient. Taken together, both provide an effective way to create a cartoon-style effect. Third, a generalization of the Kuwahara filter is presented that avoids artifacts by adapting the shape, scale, and orientation of the filter to the local structure. This causes directional image features to be better preserved and emphasized, resulting in overall sharper edges and a more feature-abiding painterly effect. In addition to the single-scale variant, a multi-scale variant is presented, which is capable of performing a highly aggressive abstraction. Fourth, a technique that builds upon the idea of combining flow-guided smoothing with shock filtering is presented, allowing for an aggressive exaggeration and an emphasis of directional image features. All presented techniques are suitable for temporally coherent per-frame filtering of video or dynamic 3D renderings, without requiring expensive extra processing, such as optical flow. Moreover, they can be efficiently implemented to process content in real-time on a GPU.
Interactive rendering techniques for focus+context visualization of 3D geovirtual environments
(2013)
This thesis introduces a collection of new real-time rendering techniques and applications for focus+context visualization of interactive 3D geovirtual environments such as virtual 3D city and landscape models. These environments are generally characterized by a large number of objects and are of high complexity with respect to geometry and textures. For these reasons, their interactive 3D rendering represents a major challenge. Their 3D depiction implies a number of weaknesses such as occlusions, cluttered image contents, and partial screen-space usage. To overcome these limitations and, thus, to facilitate the effective communication of geo-information, principles of focus+context visualization can be used for the design of real-time 3D rendering techniques for 3D geovirtual environments (see Figure). In general, detailed views of a 3D geovirtual environment are combined seamlessly with abstracted views of the context within a single image. To perform the real-time image synthesis required for interactive visualization, dedicated parallel processors (GPUs) for rasterization of computer graphics primitives are used. For this purpose, the design and implementation of appropriate data structures and rendering pipelines are necessary. The contribution of this work comprises the following five real-time rendering methods: • The rendering technique for 3D generalization lenses enables the combination of different 3D city geometries (e.g., generalized versions of a 3D city model) in a single image in real time. The method is based on a generalized and fragment-precise clipping approach, which uses a compressible, raster-based data structure. It enables the combination of detailed views in the focus area with the representation of abstracted variants in the context area. • The rendering technique for the interactive visualization of dynamic raster data in 3D geovirtual environments facilitates the rendering of 2D surface lenses. It enables a flexible combination of different raster layers (e.g., aerial images or videos) using projective texturing for decoupling image and geometry data. Thus, various overlapping and nested 2D surface lenses of different contents can be visualized interactively. • The interactive rendering technique for image-based deformation of 3D geovirtual environments enables the real-time image synthesis of non-planar projections, such as cylindrical and spherical projections, as well as multi-focal 3D fisheye-lenses and the combination of planar and non-planar projections. • The rendering technique for view-dependent multi-perspective views of 3D geovirtual environments, based on the application of global deformations to the 3D scene geometry, can be used for synthesizing interactive panorama maps to combine detailed views close to the camera (focus) with abstract views in the background (context). This approach reduces occlusions, increases the usage the available screen space, and reduces the overload of image contents. • The object-based and image-based rendering techniques for highlighting objects and focus areas inside and outside the view frustum facilitate preattentive perception. The concepts and implementations of interactive image synthesis for focus+context visualization and their selected applications enable a more effective communication of spatial information, and provide building blocks for design and development of new applications and systems in the field of 3D geovirtual environments.
In einigen Bereichen der Informatiklehre ist es möglich, die persönlichen Erfahrungen der Studierenden im Umgang mit Informationstechnik aufzugreifen und vor dem Hintergrund theoretischer Konzepte aus der Literatur gemeinsam mit ihnen zu reflektieren. Das hier vorgestellte Lehrkonzept des Reflexionsdialogs erstreckt sich über drei Seminartermine und vorbereitende Selbstlernphasen. Unterstützt wird das Konzept durch DialogueMaps, eine Software zur Visualisierung komplexer Sachverhalte und zur Unterstützung interaktiver Dialoge. Dieser Beitrag beschreibt die Hintergründe des Lehrkonzeptes, das Lehrkonzept selbst sowie die inhaltliche Ausgestaltung im Rahmen eines Mastermoduls „Computergestützte Kooperation“.
Scientific writing is an important skill for computer science and computer engineering professionals. In this paper we present a writing concept across the curriculum program directed towards scientific writing. The program is built around a hierarchy of learning outcomes. The hierarchy is constructed through analyzing the learning outcomes in relation to competencies that are needed to fulfill them.
Viele Hochschulen nutzen SAP ERP in der Lehre, um den Studierenden einen Einblick in die Funktionsweise und den Aufbau von integrierter Standardsoftware zu ermöglichen. Im Rahmen solcher Schulungen bilden die Studierenden eine Meinung und Bewertung der Software. In diesem Artikel wird untersucht, wie sich klassische Modelle der Nutzungswahrnehmung auf die spezielle Situation von SAP ERP in der Lehre übertragen lassen und welchen Einfluss bestimmte Faktoren haben. Dazu wurden vier Vorher-Nachher-Studien durchgeführt. Die Ergebnisse zeigen, dass die Funktionalität im Laufe der Schulung positiver und die Benutzungsfreundlichkeit als negativer bewertet wird.
3D from 2D touch
(2013)
While interaction with computers used to be dominated by mice and keyboards, new types of sensors now allow users to interact through touch, speech, or using their whole body in 3D space. These new interaction modalities are often referred to as "natural user interfaces" or "NUIs." While 2D NUIs have experienced major success on billions of mobile touch devices sold, 3D NUI systems have so far been unable to deliver a mobile form factor, mainly due to their use of cameras. The fact that cameras require a certain distance from the capture volume has prevented 3D NUI systems from reaching the flat form factor mobile users expect. In this dissertation, we address this issue by sensing 3D input using flat 2D sensors. The systems we present observe the input from 3D objects as 2D imprints upon physical contact. By sampling these imprints at very high resolutions, we obtain the objects' textures. In some cases, a texture uniquely identifies a biometric feature, such as the user's fingerprint. In other cases, an imprint stems from the user's clothing, such as when walking on multitouch floors. By analyzing from which part of the 3D object the 2D imprint results, we reconstruct the object's pose in 3D space. While our main contribution is a general approach to sensing 3D input on 2D sensors upon physical contact, we also demonstrate three applications of our approach. (1) We present high-accuracy touch devices that allow users to reliably touch targets that are a third of the size of those on current touch devices. We show that different users and 3D finger poses systematically affect touch sensing, which current devices perceive as random input noise. We introduce a model for touch that compensates for this systematic effect by deriving the 3D finger pose and the user's identity from each touch imprint. We then investigate this systematic effect in detail and explore how users conceptually touch targets. Our findings indicate that users aim by aligning visual features of their fingers with the target. We present a visual model for touch input that eliminates virtually all systematic effects on touch accuracy. (2) From each touch, we identify users biometrically by analyzing their fingerprints. Our prototype Fiberio integrates fingerprint scanning and a display into the same flat surface, solving a long-standing problem in human-computer interaction: secure authentication on touchscreens. Sensing 3D input and authenticating users upon touch allows Fiberio to implement a variety of applications that traditionally require the bulky setups of current 3D NUI systems. (3) To demonstrate the versatility of 3D reconstruction on larger touch surfaces, we present a high-resolution pressure-sensitive floor that resolves the texture of objects upon touch. Using the same principles as before, our system GravitySpace analyzes all imprints and identifies users based on their shoe soles, detects furniture, and enables accurate touch input using feet. By classifying all imprints, GravitySpace detects the users' body parts that are in contact with the floor and then reconstructs their 3D body poses using inverse kinematics. GravitySpace thus enables a range of applications for future 3D NUI systems based on a flat sensor, such as smart rooms in future homes. We conclude this dissertation by projecting into the future of mobile devices. Focusing on the mobility aspect of our work, we explore how NUI devices may one day augment users directly in the form of implanted devices.
This thesis proposes a privacy protection framework for the controlled distribution and use of personal private data. The framework is based on the idea that privacy policies can be set directly by the data owner and can be automatically enforced against the data user. Data privacy continues to be a very important topic, as our dependency on electronic communication maintains its current growth, and private data is shared between multiple devices, users and locations. The growing amount and the ubiquitous availability of personal private data increases the likelihood of data misuse. Early privacy protection techniques, such as anonymous email and payment systems have focused on data avoidance and anonymous use of services. They did not take into account that data sharing cannot be avoided when people participate in electronic communication scenarios that involve social interactions. This leads to a situation where data is shared widely and uncontrollably and in most cases the data owner has no control over further distribution and use of personal private data. Previous efforts to integrate privacy awareness into data processing workflows have focused on the extension of existing access control frameworks with privacy aware functions or have analysed specific individual problems such as the expressiveness of policy languages. So far, very few implementations of integrated privacy protection mechanisms exist and can be studied to prove their effectiveness for privacy protection. Second level issues that stem from practical application of the implemented mechanisms, such as usability, life-time data management and changes in trustworthiness have received very little attention so far, mainly because they require actual implementations to be studied. Most existing privacy protection schemes silently assume that it is the privilege of the data user to define the contract under which personal private data is released. Such an approach simplifies policy management and policy enforcement for the data user, but leaves the data owner with a binary decision to submit or withhold his or her personal data based on the provided policy. We wanted to empower the data owner to express his or her privacy preferences through privacy policies that follow the so-called Owner-Retained Access Control (ORAC) model. ORAC has been proposed by McCollum, et al. as an alternate access control mechanism that leaves the authority over access decisions by the originator of the data. The data owner is given control over the release policy for his or her personal data, and he or she can set permissions or restrictions according to individually perceived trust values. Such a policy needs to be expressed in a coherent way and must allow the deterministic policy evaluation by different entities. The privacy policy also needs to be communicated from the data owner to the data user, so that it can be enforced. Data and policy are stored together as a Protected Data Object that follows the Sticky Policy paradigm as defined by Mont, et al. and others. We developed a unique policy combination approach that takes usability aspects for the creation and maintenance of policies into consideration. Our privacy policy consists of three parts: A Default Policy provides basic privacy protection if no specific rules have been entered by the data owner. An Owner Policy part allows the customisation of the default policy by the data owner. And a so-called Safety Policy guarantees that the data owner cannot specify disadvantageous policies, which, for example, exclude him or her from further access to the private data. The combined evaluation of these three policy-parts yields the necessary access decision. The automatic enforcement of privacy policies in our protection framework is supported by a reference monitor implementation. We started our work with the development of a client-side protection mechanism that allows the enforcement of data-use restrictions after private data has been released to the data user. The client-side enforcement component for data-use policies is based on a modified Java Security Framework. Privacy policies are translated into corresponding Java permissions that can be automatically enforced by the Java Security Manager. When we later extended our work to implement server-side protection mechanisms, we found several drawbacks for the privacy enforcement through the Java Security Framework. We solved this problem by extending our reference monitor design to use Aspect-Oriented Programming (AOP) and the Java Reflection API to intercept data accesses in existing applications and provide a way to enforce data owner-defined privacy policies for business applications.
Cloud computing is a model for enabling on-demand access to a shared pool of computing resources. With virtually limitless on-demand resources, a cloud environment enables the hosted Internet application to quickly cope when there is an increase in the workload. However, the overhead of provisioning resources exposes the Internet application to periods of under-provisioning and performance degradation. Moreover, the performance interference, due to the consolidation in the cloud environment, complicates the performance management of the Internet applications. In this dissertation, we propose two approaches to mitigate the impact of the resources provisioning overhead. The first approach employs control theory to scale resources vertically and cope fast with workload. This approach assumes that the provider has knowledge and control over the platform running in the virtual machines (VMs), which limits it to Platform as a Service (PaaS) and Software as a Service (SaaS) providers. The second approach is a customer-side one that deals with the horizontal scalability in an Infrastructure as a Service (IaaS) model. It addresses the trade-off problem between cost and performance with a multi-goal optimization solution. This approach finds the scale thresholds that achieve the highest performance with the lowest increase in the cost. Moreover, the second approach employs a proposed time series forecasting algorithm to scale the application proactively and avoid under-utilization periods. Furthermore, to mitigate the interference impact on the Internet application performance, we developed a system which finds and eliminates the VMs suffering from performance interference. The developed system is a light-weight solution which does not imply provider involvement. To evaluate our approaches and the designed algorithms at large-scale level, we developed a simulator called (ScaleSim). In the simulator, we implemented scalability components acting as the scalability components of Amazon EC2. The current scalability implementation in Amazon EC2 is used as a reference point for evaluating the improvement in the scalable application performance. ScaleSim is fed with realistic models of the RUBiS benchmark extracted from the real environment. The workload is generated from the access logs of the 1998 world cup website. The results show that optimizing the scalability thresholds and adopting proactive scalability can mitigate 88% of the resources provisioning overhead impact with only a 9% increase in the cost.
Learning a model for the relationship between the attributes and the annotated labels of data examples serves two purposes. Firstly, it enables the prediction of the label for examples without annotation. Secondly, the parameters of the model can provide useful insights into the structure of the data. If the data has an inherent partitioned structure, it is natural to mirror this structure in the model. Such mixture models predict by combining the individual predictions generated by the mixture components which correspond to the partitions in the data. Often the partitioned structure is latent, and has to be inferred when learning the mixture model. Directly evaluating the accuracy of the inferred partition structure is, in many cases, impossible because the ground truth cannot be obtained for comparison. However it can be assessed indirectly by measuring the prediction accuracy of the mixture model that arises from it. This thesis addresses the interplay between the improvement of predictive accuracy by uncovering latent cluster structure in data, and further addresses the validation of the estimated structure by measuring the accuracy of the resulting predictive model. In the application of filtering unsolicited emails, the emails in the training set are latently clustered into advertisement campaigns. Uncovering this latent structure allows filtering of future emails with very low false positive rates. In order to model the cluster structure, a Bayesian clustering model for dependent binary features is developed in this thesis. Knowing the clustering of emails into campaigns can also aid in uncovering which emails have been sent on behalf of the same network of captured hosts, so-called botnets. This association of emails to networks is another layer of latent clustering. Uncovering this latent structure allows service providers to further increase the accuracy of email filtering and to effectively defend against distributed denial-of-service attacks. To this end, a discriminative clustering model is derived in this thesis that is based on the graph of observed emails. The partitionings inferred using this model are evaluated through their capacity to predict the campaigns of new emails. Furthermore, when classifying the content of emails, statistical information about the sending server can be valuable. Learning a model that is able to make use of it requires training data that includes server statistics. In order to also use training data where the server statistics are missing, a model that is a mixture over potentially all substitutions thereof is developed. Another application is to predict the navigation behavior of the users of a website. Here, there is no a priori partitioning of the users into clusters, but to understand different usage scenarios and design different layouts for them, imposing a partitioning is necessary. The presented approach simultaneously optimizes the discriminative as well as the predictive power of the clusters. Each model is evaluated on real-world data and compared to baseline methods. The results show that explicitly modeling the assumptions about the latent cluster structure leads to improved predictions compared to the baselines. It is beneficial to incorporate a small number of hyperparameters that can be tuned to yield the best predictions in cases where the prediction accuracy can not be optimized directly.
Im Bachelor-Studiengang (B. Sc.) IT Security an der Fachhochschule St. Pölten wurde im Wintersemester 2011/12 versuchsweise die Lehrorganisation im ersten Fachsemester verändert: Die Module bzw. Teilmodule wurden nicht mehr alle parallel zueinander unterrichtet, sondern jedes Modul wurde exklusiv über einige Wochen abgehalten. Im Beitrag werden die Auswirkungen und bisherigen Erfahrungen mit dieser Reorganisation der Lehre geschildert: So haben sich die Noten im Mittel um etwa eine Note verbessert, die Zahl derjenigen Studierenden, die durch Prüfungen durchfallen, ist drastisch gesunken. Die Zufriedenheit der Studierenden und Lehrenden ist so groß, dass diese Form der Lehrorganisation im gesamten Bachelor- und auch im Masterstudiengang übernommen wird.
Der vorliegende Beitrag beschäftigt sich mit der Frage, wie der eLearning-Support in großen Institutionen effizient gestaltet werden kann. Vorgestellt wird ein experimentelles Beratungsprojekt, das Lehrende bei der Gestaltung von eLearning-Maßnahmen mithilfe der Lernplattform ILIAS1 unterstützt. Neben der Zielsetzung des Projekts werden dessen Aufbau und erste Praxiserfahrungen erörtert. Außerdem werden Potenziale des Beratungsformats, die insbesondere mit der individuellen Vor-Ort-Beratung der Lehrenden durch hochschuldidaktisch geschulte Studierende einhergehen, erläutert. Abschließend werden Grenzen und Weiterentwicklungsperspektiven des Projekts dargestellt. Am Beispiel der ILIAS-Beratung soll gezeigt werden, dass es sich einer nachhaltigen Organisationsentwicklung als zuträglich erweist, Kooperationen erschiedenartiger Organisationseinheiten zu fördern und die entstehenden Synergieeffekte zu nutzen.
Wir stellen die Konzeption und erste Ergebnisse einer neuartigen Informatik- Lehrveranstaltung für Studierende der Geodäsie vor. Das Konzept verbindet drei didaktische Ideen: Kontextorientierung, Peer-Tutoring und Praxisbezug (Course). Die Studierenden sollen dabei in zwei Semestern wichtige Grundlagen der Informatik verstehen und anzuwenden lernen. Durch enge Verzahnung der Aufgaben mit einem für Nichtinformatiker relevanten Kontext, sowie einem sehr hohen Anteil von Selbsttätigkeit der Studierenden soll die Motivation für fachfremde Themen gesteigert werden. Die Ergebnisse zeigen, dass die Veranstaltung sehr erfolgreich war.
In diesem Beitrag berichten wir über die Erfahrungen einer umgestalteten Lehre im Bereich Informatik und Gesellschft (IuG). Die Gründe für die Umge staltung und die Konzeption werden skizziert. Die Erfahrungen haben wir zu Thesen verdichtet: 1. Informatik und Gesellschaft sollte eine Pflichtveranstaltung im Bachelor-Studium sein, in der Studierende einen Überblick erhalten, welche gesellschaftlichen Rahmenbedingungen für sie relevant sind und wie man diese in die Praxis mit einbeziehen kann. 2. Historische Inhalte der Informatik sollen hier aufgearbeitet werden, indem man aktuelle Entwicklungen im Kontext ihrer Genese betrachtet.
Zurzeit haben wir es mit der folgenden Situation an Universitäten zu tun: Studierende kommen mit unterschiedlichem Wissen und Kompetenzen zur Universität, um informatikbezogene Studiengänge zu belegen. Diesem Umstand muss in den universitären Kursen entgegengewirkt werden, um ein einheitliches Bildungsziel zu erreichen. Für einige Studierende bedeutet dies oft eine Lehrbelastung in einem ohnehin sehr zeitintensiven Studium, was nicht selten zum Studienabbruch führt. Ein anderes Problem ist die fehlende Transparenz bezüglich der Gegenstände des Informatikstudiums: einige angehende Studierende kommen mit einem von der Realität abweichenden Bild der Informatik zur Universität, andere entscheiden sich u. U. deshalb gegen ein Informatikstudium, da ihnen nicht bewusst ist, dass das Studium für sie interessant sein könnte. In diesem Artikel soll ein Lösungsvorschlag anhand eines Kompetenzrahmenmodells vorgestellt werden, mit dessen Hilfe eine Verbesserung der Hochschulsituation erreicht werden kann.
A survey has been carried out in the Computer Science (CS) department at the University of Baghdad to investigate the attitudes of CS students in a female dominant environment, showing the differences between male and female students in different academic years. We also compare the attitudes of the freshman students of two different cultures (University of Baghdad, Iraq, and the University of Potsdam).
This talk will describe My Digital Life (TU100), a distance learning module that introduces computer science through immediate engagement with ubiquitous computing (ubicomp). This talk will describe some of the principles and concepts we have adopted for this modern computing introduction: the idea of the ‘informed digital citizen’; engagement through narrative; playful pedagogy; making the power of ubicomp available to novices; setting technical skills in real contexts. It will also trace how the pedagogy is informed by experiences and research in Computer Science education.
Die Orientierung am Outcome eines Lernprozesses stellt einen wichtigen Pfeiler einer kompetenzorientierten Informatiklehre dar. Im Beitrag werden Konzeption und Erfahrungen eines Projekts zur outcome-orientierten Neuausrichtung der Informatiklehre unter Berücksichtigung der Theorie des Constructive Alignment beschrieben. Nach der theoretischen Fundierung der Kompetenzproblematik wird anhand eines Formulierungsmodells ein Prozess zur Erarbeitung beobachtbarer Lernergebnisse dargestellt. Die Diskussion der Projektziele und Erfahrungen in der Umsetzung und Evaluierung unterstreichen die Chancen und Herausforderungen für eine Steigerung der Studienqualität.
Zur Unterstützung von Studierenden in der Studieneingangsphase wurde an der RWTH Aachen ein neuartiger und motivierender Einstieg in den Vorkurs Informatik entwickelt und zum Wintersemester 2011/12 erprobt. Dabei wurde die grafische Programmierung mittels App Inventor eingeführt, die zur Umsetzung anwendungsbezogener Projekte genutzt wurde. In diesem Beitrag werden die Motivation für die Neugestaltung, das Konzept und die Evaluation des Testlaufs beschrieben. Diese dienen als Grundlage für eine vollständige Neukonzeption des Vorkurses für das Wintersemester 2012/2013.
In many applications one is faced with the problem of inferring some functional relation between input and output variables from given data. Consider, for instance, the task of email spam filtering where one seeks to find a model which automatically assigns new, previously unseen emails to class spam or non-spam. Building such a predictive model based on observed training inputs (e.g., emails) with corresponding outputs (e.g., spam labels) is a major goal of machine learning. Many learning methods assume that these training data are governed by the same distribution as the test data which the predictive model will be exposed to at application time. That assumption is violated when the test data are generated in response to the presence of a predictive model. This becomes apparent, for instance, in the above example of email spam filtering. Here, email service providers employ spam filters and spam senders engineer campaign templates such as to achieve a high rate of successful deliveries despite any filters. Most of the existing work casts such situations as learning robust models which are unsusceptible against small changes of the data generation process. The models are constructed under the worst-case assumption that these changes are performed such to produce the highest possible adverse effect on the performance of the predictive model. However, this approach is not capable to realistically model the true dependency between the model-building process and the process of generating future data. We therefore establish the concept of prediction games: We model the interaction between a learner, who builds the predictive model, and a data generator, who controls the process of data generation, as an one-shot game. The game-theoretic framework enables us to explicitly model the players' interests, their possible actions, their level of knowledge about each other, and the order at which they decide for an action. We model the players' interests as minimizing their own cost function which both depend on both players' actions. The learner's action is to choose the model parameters and the data generator's action is to perturbate the training data which reflects the modification of the data generation process with respect to the past data. We extensively study three instances of prediction games which differ regarding the order in which the players decide for their action. We first assume that both player choose their actions simultaneously, that is, without the knowledge of their opponent's decision. We identify conditions under which this Nash prediction game has a meaningful solution, that is, a unique Nash equilibrium, and derive algorithms that find the equilibrial prediction model. As a second case, we consider a data generator who is potentially fully informed about the move of the learner. This setting establishes a Stackelberg competition. We derive a relaxed optimization criterion to determine the solution of this game and show that this Stackelberg prediction game generalizes existing prediction models. Finally, we study the setting where the learner observes the data generator's action, that is, the (unlabeled) test data, before building the predictive model. As the test data and the training data may be governed by differing probability distributions, this scenario reduces to learning under covariate shift. We derive a new integrated as well as a two-stage method to account for this data set shift. In case studies on email spam filtering we empirically explore properties of all derived models as well as several existing baseline methods. We show that spam filters resulting from the Nash prediction game as well as the Stackelberg prediction game in the majority of cases outperform other existing baseline methods.