Institut für Informatik und Computational Science
Refine
Year of publication
- 2023 (1)
- 2022 (15)
- 2021 (14)
- 2020 (22)
- 2019 (10)
- 2018 (21)
- 2017 (4)
- 2016 (9)
- 2015 (63)
- 2014 (29)
- 2013 (37)
- 2012 (23)
- 2011 (27)
- 2010 (32)
- 2009 (26)
- 2008 (1)
- 2007 (24)
- 2006 (38)
- 2005 (29)
- 2004 (27)
- 2003 (18)
- 2002 (15)
- 2001 (20)
- 2000 (31)
- 1999 (29)
- 1998 (30)
- 1997 (22)
- 1996 (25)
- 1995 (15)
- 1994 (17)
- 1993 (1)
Document Type
- Article (676) (remove)
Keywords
- Didaktik (13)
- Ausbildung (12)
- Hochschuldidaktik (12)
- Informatik (12)
- Answer set programming (10)
- answer set programming (8)
- Answer Set Programming (6)
- Computer Science Education (4)
- E-Learning (4)
- Competence Measurement (3)
- Machine learning (3)
- Secondary Education (3)
- Theory (3)
- formal languages (3)
- monitoring (3)
- Analytical models (2)
- Automata systems (2)
- Big Data (2)
- Competence Modelling (2)
- Computational thinking (2)
- Digitale Medien (2)
- E-learning (2)
- Equilibrium logic (2)
- Event mapping (2)
- Fault tolerance (2)
- IT-Infrastruktur (2)
- Informatics Education (2)
- Informatics Modelling (2)
- Informatics System Application (2)
- Informatics System Comprehension (2)
- Lindenmayer systems (2)
- Non-monotonic reasoning (2)
- Parameterized complexity (2)
- Process mining (2)
- ResNet (2)
- Strategie (2)
- bioinformatics (2)
- computational thinking (2)
- computer science education (2)
- cooperating systems (2)
- e-learning (2)
- education (2)
- higher education (2)
- online learning (2)
- radhard design (2)
- relevance (2)
- security (2)
- tools (2)
- (FPGA) (1)
- (SET) count rate (1)
- 2-tag system (1)
- 21st century skills, (1)
- 3D modeling (1)
- 3D visualization (1)
- ABRACADABRA (1)
- AODV (1)
- ARCS Modell (1)
- ASIC (1)
- Abstraction (1)
- Access control (1)
- Achievement (1)
- Active evaluation (1)
- Activity Theory (1)
- Activity-orientated Learning (1)
- Ad hoc routing (1)
- Adaptivity (1)
- Adaptivität (1)
- Advanced Video Codec (AVC) (1)
- Aggregates (1)
- Algorithm configuration (1)
- Algorithm portfolios (1)
- Algorithms (1)
- Android hybrid apps (1)
- Animal building (1)
- Anti-cancer drugs (1)
- Arduino (1)
- Argumentation structure (1)
- Assessment (1)
- Audience Response Systeme (1)
- Augmentation (1)
- Augmented and virtual reality (1)
- Austria (1)
- Autismus (1)
- Automated Theorem Proving (1)
- Automated parallelization (1)
- Automatic Item Generation (1)
- Automatically controlled windows (1)
- Automatisches Beweisen (1)
- Backdoors (1)
- Batch processing (1)
- Bayesian inference (1)
- Benchmark testing; (1)
- Berliner Modell (1)
- Blended Learning (1)
- Blind users (1)
- Bloom’s Taxonomy (1)
- Boolean logic models (1)
- Business process intelligence (1)
- CP-Logic (1)
- CS concepts (1)
- Campus (1)
- Capability approach (1)
- Challenges (1)
- Circuit faults (1)
- Clause Learning (1)
- Clock tree (1)
- Cloud (1)
- Cluster Computing (1)
- Cluster computing (1)
- Code generation (1)
- Cognitive Skills (1)
- Coherent phonons (1)
- Combinatorial multi-objective optimization (1)
- Competences (1)
- Competencies (1)
- Complex optimization (1)
- Complexity (1)
- Computational Thinking (1)
- Computational complexity (1)
- Computational grid (1)
- Computer Science (1)
- Computer Science in Context (1)
- Computer security (1)
- Computergestützes Training (1)
- Computing (1)
- Computing with DNA (1)
- Conformant Planning (1)
- Conrad Hal Waddington (1)
- Constraint satisfaction (1)
- Contest (1)
- Context awareness (1)
- Contextualisation (1)
- Contextualized learning (1)
- Continuous Testing (1)
- Continuous Versioning (1)
- Contradictions (1)
- Convolution (1)
- Course timetabling (1)
- Curriculum (1)
- Curriculum Development (1)
- Customer ownership (1)
- D-galactosamine (1)
- DMR (1)
- DNA hairpin formation (1)
- DPLL (1)
- DRMAA (1)
- DRMS (1)
- Data Analysis (1)
- Data Management (1)
- Data Privacy (1)
- Data augmentation (1)
- Data federation (1)
- Database (1)
- Databases (1)
- Deal of the Day (1)
- Debugging (1)
- Decidability (1)
- Declare (1)
- Deep learning (1)
- Defining characteristics of physical computing (1)
- Denotational semantics (1)
- Design (1)
- Design for testability (DFT) (1)
- Digital Competence (1)
- Digital Education (1)
- Digital Game Based Learning (1)
- Digital Revolution (1)
- Digitalisierung (1)
- Digitalization (1)
- Double cell upsets (DCUs) (1)
- Dynamic assessment (1)
- Dynamical X-ray theory (1)
- E-Assessment (1)
- E-Klausuren (1)
- E-Portfolio (1)
- E-teaching (1)
- EDC (1)
- EEG (1)
- Early Literacy (1)
- Earthquake modeling (1)
- Educational Standards (1)
- Educational game (1)
- Educational software (1)
- Educational timetabling (1)
- Elektronisches Prüfen (1)
- Embedded Systems (1)
- Encoding (1)
- Engines (1)
- Entity Linking (1)
- Epigenetic landscape (1)
- Epistemic Logic Programs (1)
- Euclid’s algorithm (1)
- Evaluation (1)
- Evolution (1)
- Experimentation (1)
- Explicit negation (1)
- Explore-first Programming (1)
- Extensibility (1)
- Extreme Model-Driven Development (1)
- FEDC (1)
- FPGA (1)
- Facebook (1)
- Fault Localization (1)
- Fault tolerant systems (1)
- Feature extraction (1)
- Fibonacci numbers (1)
- Field programmable gate arrays (1)
- Finite automata (1)
- Flip-flops (1)
- Forgetting (1)
- Forschungsdatenmanagement (1)
- Freshmen (1)
- Function (1)
- Fundamental Ideas (1)
- GERBIL (1)
- Gaussian process (1)
- Gesture input (1)
- Graphensuche (1)
- Green computing (1)
- Grounded theory (1)
- H.264 (1)
- Hairpin completions (1)
- Hairpin reductions (1)
- Hardware accelerator (1)
- Hawkes process (1)
- Heat diffusion (1)
- Hierarchically configurable mask register (1)
- Histograms (1)
- Hochschul-Cloud (1)
- Hochschullehre (1)
- Https traffic (1)
- Human Factors (1)
- Hurricane Sandy (1)
- ICT Competence (1)
- ICT competencies (1)
- ICT skills (1)
- IaaS (1)
- Identifiers (1)
- Image and video stylization (1)
- Image resolution (1)
- Imperative calculi (1)
- Improving classroom (1)
- Incoherent phonons (1)
- Incremental answer set programming (1)
- Inference (1)
- Informatics (1)
- Informatikstudium (1)
- Information federation (1)
- Information integration (1)
- Information retrieval (1)
- Information security (1)
- Innovation (1)
- Inquiry-based Learning (1)
- Insurance industry (1)
- Integrated circuit modeling (1)
- Interaktivität (1)
- Interface design (1)
- Internet of Things (1)
- Job monitoring (1)
- Job submission (1)
- Kernel (1)
- Kernelization (1)
- Key Competencies (1)
- Key input (1)
- Klausellernen (1)
- Knowledge representation (1)
- Kompetenzerwerb (1)
- L systems (1)
- LBA problem (1)
- LMS (1)
- Landmark visibility (1)
- Learners (1)
- Learning Fields (1)
- Learning ecology (1)
- Learning interfaces development (1)
- Learning with ICT (1)
- Lehrer*innenbildung (1)
- Lern-App (1)
- Lernaufgaben (1)
- Lernmotivation (1)
- Literature mining (1)
- Liver neoplasms (1)
- Load Balancing (1)
- Localization (1)
- Location awareness (1)
- Logarithm (1)
- Logic programming (1)
- Low Latency (1)
- Lower Secondary Level (1)
- Loyalty (1)
- MOOCs (1)
- MQTT (1)
- Machine Learning (1)
- Markov processes (1)
- Masking of X-values (1)
- Massive Open Online Courses (1)
- Measurement (1)
- Media in education (1)
- Meta-Programming (1)
- Metric learning (1)
- Minimal perturbation problems (1)
- Mobile App (1)
- Mobile application (1)
- Mobile devices (1)
- Mobile learning (1)
- Model checking (1)
- Modeling (1)
- Modelling (1)
- Multi-objective optimization (1)
- Multi-sided platforms (1)
- Multiple interpretation scheme (1)
- Music Technology (1)
- N-temperature model (1)
- NUI (1)
- Nash equilibrium (1)
- Natural Science Education (1)
- Natural language processing (1)
- Natural ventilation (1)
- Network (1)
- Network security (1)
- Neural networks (1)
- NoSQL (1)
- Non-Monotonic (1)
- Nonmonotonic reasoning (1)
- Norway (1)
- Novice programmers (1)
- OCCI (1)
- OSSE (1)
- OpenOLAT (1)
- Operation problem (1)
- Optimization (1)
- Parallel SAT solving (1)
- Parallel job execution time estimation (1)
- Particle detector (1)
- Partizipation (1)
- Pedagogical content knowledge (1)
- Pedagogical issues (1)
- Pedestrian navigation (1)
- Performance Evaluation (1)
- Personalization (1)
- Persönliche Lernumgebung (1)
- Pervasive computing (1)
- Pervasive game (1)
- Pervasive learning (1)
- Phantoms (1)
- Physical Science (1)
- Planar tactile display (1)
- Polarization (1)
- Preference Handling (1)
- Preprocessing (1)
- Problem Solving (1)
- Process model analysis (1)
- Product lifecycle management (1)
- Programming (1)
- Programming by optimization (1)
- Prototyping (1)
- REST (1)
- RSA triangle (1)
- Random access memory (1)
- Ranking (1)
- Reasoning (1)
- Recommendations for CS-Curricula in Higher Education (1)
- Region of Interest (1)
- Relevanz (1)
- Reproducibility of results (1)
- Reversibility (1)
- SAT (1)
- SET pulsewidth distribution (1)
- SOA (1)
- SWOT (1)
- SaaSAbstract (1)
- Sampling (1)
- Scalability (1)
- Scale-invariant feature transform (SIFT) (1)
- Scientific images (1)
- Screen reader (1)
- Seamless learning (1)
- Search problems (1)
- Security (1)
- Self-exciting point process (1)
- Semantic data (1)
- Semantic web (1)
- Semilinearity property (1)
- Sensors (1)
- Sequence embeddings (1)
- Service orientation (1)
- Sharing (1)
- Signal processing (1)
- Signaling transduction networks (1)
- Simulations (1)
- Single event effect (1)
- Single event upsets (1)
- Single-event transient (SET) (1)
- Small Private Online Courses (1)
- Social (1)
- Spatio-temporal ETAS model (1)
- Splicing (1)
- Splicing processor (1)
- Statistical relational learning (1)
- Stochastic relational process (1)
- Strong equivalence (1)
- Structural equation modeling (1)
- Studentenjobs (1)
- Studienabbrecher (1)
- Studiendauer (1)
- Systems biology (1)
- Systems of parallel communicating (1)
- TMR (1)
- TPACK (1)
- Tasks (1)
- Teacher perceptions (1)
- Teachers (1)
- Teaching information security (1)
- Technology proficiency (1)
- Terminology (1)
- Tests (1)
- Theorembeweisen (1)
- Theory formation (1)
- Thermoelasticity (1)
- Time series (1)
- Tomography (1)
- Tools (1)
- Tracking (1)
- Traffic data (1)
- Tree decomposition (1)
- Treewidth (1)
- Treewidth-aware reductions (1)
- Triple modular redundancy (TMR) (1)
- Tumor types (1)
- Turing machine (1)
- Type and effect systems (1)
- UAV imagery (1)
- UX (1)
- Ubiquitous learning (1)
- Ultrafast dynamics (1)
- Unary languages (1)
- Unifikation (1)
- Uniform Access Principle (1)
- Usability testing (1)
- User Experience (1)
- User submission pattern (1)
- User-centred design (1)
- Value network (1)
- Verification (1)
- Visual metaphor (1)
- Vocational Education (1)
- Weiterbildung (1)
- Wireless Sensor Networks (1)
- Word embeddings (1)
- X-masking (1)
- X-ray computed (1)
- X-values (1)
- Young People (1)
- abstraction (1)
- accepting grammars (1)
- action and change (1)
- activity (1)
- acute liver failure (1)
- acyclicity properties (1)
- adversarial classification (1)
- algorithm schedules (1)
- algorithms (1)
- analogical thinking (1)
- analysis (1)
- anti-cancer drugs (1)
- anxiety (1)
- approximate model counting (1)
- architecture (1)
- argument mining (1)
- arousal (1)
- artistic rendering (1)
- asynchrounous design (1)
- autism (1)
- automata (1)
- automated guided vehicle routing (1)
- automated planning (1)
- automatic feedback (1)
- behavioral abstraction (1)
- belief merging (1)
- belief revision (1)
- benchmark (1)
- bibliometric analysis (1)
- binary representation (1)
- binary search (1)
- block representation (1)
- bootstrapping (1)
- brain-computer interface (1)
- bundled data (1)
- camera sensor (1)
- car assembly operations (1)
- cellular automata (1)
- circuit Faults (1)
- citation analysis (1)
- classroom language (1)
- click controller (1)
- clocks (1)
- co-citation analysis (1)
- co-occurrence analysis (1)
- code generation (1)
- cognitive modifiability (1)
- coherence relation (1)
- collaborative learning (1)
- combinatorial optimization problems (1)
- combined task and motion planning (1)
- common spatial patterns (1)
- competence (1)
- competencies (1)
- competency (1)
- competition (1)
- complexity (1)
- compliance (1)
- comprehension (1)
- computer science teachers (1)
- computer vision (1)
- concession (1)
- concurrent checking (1)
- conductive argument (1)
- connective (1)
- consistency (1)
- consistency checking (1)
- consistency measures (1)
- context-free grammar (1)
- context-sensitive (1)
- contrast (1)
- controlled vocabularies (1)
- corpus analysis (1)
- correlated errors (1)
- course timetabling (1)
- craters (1)
- crop (1)
- cs4fn (1)
- curriculum theory (1)
- decidability questions (1)
- declarative problem solving (1)
- deep learning (1)
- deep neural networks (1)
- deep residual networks (1)
- degree of non-context-freeness (1)
- degree of non-regularity (1)
- degree of non-regulation (1)
- depression (1)
- design flow (1)
- determinism (1)
- detrending (1)
- developmental systems (1)
- diagnosis (1)
- didactics (1)
- didaktisches Konzept (1)
- digitale Bildung (1)
- digitale Medien (1)
- digitally-enabled pedagogies (1)
- divide and conquer (1)
- domain-specific APIs (1)
- drug discovery (1)
- drug-sensitivity prediction (1)
- dynamic service binding (1)
- e-mentoring (1)
- eLectures (1)
- economic ripples (1)
- edge computing (1)
- education and public policy (1)
- educational programming (1)
- educational systems (1)
- educational timetabling (1)
- edutainment (1)
- embedded systems (1)
- emission factor (1)
- endothelin (1)
- endothelin-converting enzyme (1)
- ensemble kalman filter (1)
- ensemble methods (1)
- environments (1)
- error propagation (1)
- evaluation (1)
- event-related desynchronization (1)
- evolution (1)
- exponentiation (1)
- external ambiguity (1)
- extreme weather (1)
- face tracking (1)
- facial expression (1)
- fault tolerance (1)
- field-programmable gate array (1)
- finite model computation (1)
- finite state sequential transducers (1)
- firmware update (1)
- formal (1)
- formal argumentation systems (1)
- fun (1)
- gap-filling (1)
- geovisualization (1)
- gradient boosting (1)
- grammar (1)
- graph-search (1)
- greenhouse gas (1)
- hardware accelerator (1)
- hardware architecture (1)
- high school (1)
- higher (1)
- hybrid solving (1)
- ice harboring (1)
- image classification (1)
- image processing (1)
- image recognition (1)
- imaging (1)
- impacts (1)
- incremental SVM (1)
- informal and formal learning (1)
- informal logic (1)
- informatics (1)
- informatics education (1)
- information flow control (1)
- innovation (1)
- interactive course (1)
- interactive workshop (1)
- internal ambiguity (1)
- intrusion detection (1)
- key competences in physical computing (1)
- key competencies (1)
- kidney cancer (1)
- kinaesthetic teaching (1)
- klinisch-praktischer Unterricht (1)
- knowledge representation and nonmonotonic reasoning (1)
- knowledge representation and reasoning (1)
- latches (1)
- learning (1)
- leftmost derivations (1)
- lesson planning (1)
- lesson preparation (1)
- linear programming (1)
- logic programming (1)
- logic-based modeling (1)
- loop formulas (1)
- loose programming (1)
- loss propagation (1)
- lunar exploration (1)
- machine learning (1)
- machine learning algorithms (1)
- manipulation planning (1)
- measure development (1)
- media (1)
- mediated learning experience (1)
- metabolic network (1)
- metabolism (1)
- metabolomics (1)
- metadata (1)
- metastasis (1)
- mobile learning (1)
- mobile technologies and apps (1)
- mobiles Lernen (1)
- natural disasters (1)
- natural language generation (1)
- neighborhood (1)
- networks (1)
- neural networks (1)
- neutral endopeptidase (1)
- nonphotorealistic rendering (NPR) (1)
- o-ambiguity (1)
- on-farm evaluation (1)
- open learning (1)
- operating system (1)
- organisational evolution (1)
- paper prototyping (1)
- parallel processing (1)
- parallel rewriting (1)
- parameter (1)
- parity aggregate operator (1)
- parsing (1)
- pdf forms (1)
- pedagogy (1)
- perception (1)
- perception differences (1)
- personal (1)
- personal response systems (1)
- philosophical foundation of informatics pedagogy (1)
- physical computing (1)
- physical computing tools (1)
- planning (1)
- plug-ins (1)
- policy evaluation (1)
- portfolio-based solving (1)
- pre-primary level (1)
- predictive models (1)
- premise acceptability (1)
- preprocessing (1)
- primary education (1)
- primary level (1)
- problem-solving (1)
- process model alignment (1)
- process modeling (1)
- professional development (1)
- program encodings (1)
- programmed grammars (1)
- programming (1)
- programming in context (1)
- projection (1)
- proof complexity (1)
- pruritus (1)
- pulse stretching inverters (1)
- quality of life (1)
- quantum (1)
- random forest (1)
- real arguments (1)
- real-time (1)
- real-time mapping (1)
- reference (1)
- referential effectiveness (1)
- regression (1)
- regular language (1)
- reliability (1)
- reliability analysis (1)
- resources (1)
- restricted parallelism (1)
- satisfiability (1)
- secondary computer science education (1)
- secondary education (1)
- selective fault tolerance (1)
- self-adaptive multiprocessing system (1)
- self-checking (1)
- self-efficacy (1)
- semantic web (1)
- simplicity (1)
- single event upset (1)
- single event upsets (1)
- single-event transient (1)
- single-trial-analysis (1)
- site-specific weed management (1)
- sleep quality (1)
- smart farming (1)
- social media (1)
- soft errors (1)
- solar particle event (1)
- space missions (1)
- stable model semantics (1)
- state complexity (1)
- static analysis (1)
- static prediction games (1)
- strong equivalence (1)
- student activation (1)
- student experience (1)
- student perceptions (1)
- students’ conceptions (1)
- students’ knowledge (1)
- sufficiency (1)
- suicidal ideations (1)
- supply chains (1)
- support system (1)
- support vector machines (1)
- tableau calculi (1)
- teacher competencies (1)
- teacher training (1)
- teaching (1)
- teaching informatics in general education (1)
- technical notes and rapid communications (1)
- technische Rahmenbedingungen (1)
- tele-teaching (1)
- test response compaction (1)
- theorem (1)
- theory of computation (1)
- timing (1)
- tracing (1)
- transient Faults (1)
- transient analysis (1)
- triangulated irregular networks (1)
- triple modular redundancy (1)
- unfounded sets (1)
- unification (1)
- user experience (1)
- user-centred (1)
- verification (1)
- video annotation (1)
- virtual mobility (1)
- virtual reality (1)
- wheat crops (1)
- work productivity (1)
- xAPI (1)
- yellow rust (1)
- ‘unplugged’ computing (1)
Institute
- Institut für Informatik und Computational Science (676)
- Institut für Physik und Astronomie (2)
- eLiS - E-Learning in Studienbereichen (2)
- Department Erziehungswissenschaft (1)
- Department Linguistik (1)
- Extern (1)
- Historisches Institut (1)
- Universitätsbibliothek (1)
- Zentrum für Qualitätsentwicklung in Lehre und Studium (ZfQ) (1)
Das „Startprojekt“
(2016)
Absolventinnen und Absolventen unserer Informatik-Bachelorstudiengänge benötigen für kompetentes berufliches Handeln sowohl fachliche als auch überfachliche Kompetenzen. Vielfach verlangen wir von Erstsemestern in Grundlagen-Lehrveranstaltungen fast ausschließlich den Aufbau von Fachkompetenz und vernachlässigen dabei häufig Selbstkompetenz, Methodenkompetenz und Sozialkompetenz. Gerade die drei letztgenannten sind für ein erfolgreiches Studium unabdingbar und sollten von Anfang an entwickelt werden. Wir stellen unser „Startprojekt“ als einen Beitrag vor, im ersten Semester die eigenverantwortliche, überfachliche Kompetenzentwicklung in einem fachlichen Kontext zu fördern.
Aus einer Vergleichsstudie mit starken und schwachen Problemlösern konnten Erkenntnisse über die effizienten Herangehensweisen von Hochleistern an Informatikprobleme gewonnen werden. Diese Erkenntnisse wurden in einem Lehrvideo zum informatischen Problemlösen didaktisch aufgearbeitet, sodass Lernenden der Einsatz von Baumstrukturen und Rekursion im konkreten Kontext gezeigt werden kann. Nun wurde die tatsächliche Lernwirksamkeit des Videos sowie die Definition der Zielgruppe in einer Vergleichsstudie mit 66 Studienanfängern überprüft.
Der Unterricht großer Studierendengruppen im wissenschaftlichen Schreiben birgt vielfältige organisatorische Herausforderungen und eine zeitintensive Betreuung durch die Dozenten. Diese Arbeit stellt ein Lehrkonzept mit Peer-Reviews vor, in dem das Feedback der Peers durch eine automatisierte Analyse ergänzt wird. Die Software Confopy liefert metrik- und strukturbasierte Hinweise für die Verbesserung des wissenschaftlichen Schreibstils. Der Nutzen von Confopy wird an 47 studentischen Arbeiten in Draft- und Final-Version illustriert.
Informatik-Studierende haben in der Mehrzahl Schwierigkeiten, einen Einstieg in die Theoretische
Informatik zu finden und die Leistungsanforderungen in den
Endklausuren der zugehörigen Lehrveranstaltungen zu erfüllen. Wir argumentieren, dass dieser Symptomatik mangelnde Kompetenzen im Umgang mit abstrakten und stark formalisierten Themeninhalten zugrunde liegen und schlagen vor, einen Beweisassistenten als interaktives Lernwerkzeug in der Eingangslehre der Theoretischen Informatik zu nutzen, um entsprechende Kompetenzen zu stärken.
Dieser Beitrag diskutiert den Einsatz von interaktiven und automatischen Theorembeweisern in der universitären Lehre. Moderne Theorembeweiser scheinen geeignet zur Implementierung des dialogischen Lernens und als E-Assessment-Werkzeug in der Logikausbilding. Exemplarisch skizzieren wir ein innovaties Lehrprojekt zum Thema „Komputationale Metaphysik“, in dem die zuvor genannten Werkzeuge eingesetzt werden.
Die Unterrichtsmethode Stationsarbeit kann verwendet werden, um Individualisierung und Differenzierung im Lernprozess zu ermöglichen. Dieser Beitrag schlägt Aufgabenformate vor, die in einer Stationsarbeit über das Klassendiagramm aus der Unified Modeling Language verwendet werden können. Die Aufgabenformate wurden bereits mit Studierenden erprobt.
Solving problems combining task and motion planning requires searching across a symbolic search space and a geometric search space. Because of the semantic gap between symbolic and geometric representations, symbolic sequences of actions are not guaranteed to be geometrically feasible. This compels us to search in the combined search space, in which frequent backtracks between symbolic and geometric levels make the search inefficient.We address this problem by guiding symbolic search with rich information extracted from the geometric level through culprit detection mechanisms.
Die Vermittlung von Modellierungsfähigkeiten in der Softwaretechnik-Ausbildung konzentriert sich meist auf Modellierungskonzepte, Notationen und Entwicklungswerkzeuge. Die Betrachtung der Modellierungsaktivitäten, etwa die Entwicklung und Gegenüberstellung alternativer Modellvorschläge, steht weniger im Vordergrund. Die vorliegende Studie untersucht zwei Formen des kollaborativen Modellierens am Tabletop in Bezug auf ihren Einfluss auf die Modellierungsaktivitäten in kleinen Gruppen. Die Ergebnisse zeigen, dass sowohl selbstorganisierte als auch moderierte Modellierungssitzungen das Entwickeln eines gemeinsamen Modellverständnisses fördern. In moderierten Sitzungen wurden zudem mehr alternative Lösungsideen entwickelt und in stärkerem Maße diskutiert.
Recent philosophical analyses of the epistemic dimension of images in the sciences show a certain trend in acknowledging potential roles of these images beyond their merely decorative or pedagogical functions. We argue, however, that this new debate has yet paid little attention to a special type of pictures, we call ‘visual metaphor’, and its versatile heuristic potential in organizing data, supporting communication, and guiding research, modeling, and theory formation. Based on a case study of Conrad Hal Waddington’s epigenetic landscape images in biology, we develop a descriptive framework applicable to heuristic roles of various visual metaphors in the sciences.
Answer Set Programming (ASP) is a well-known paradigm of declarative programming with roots in logic programming and non-monotonic reasoning. Similar to other closely related problemsolving technologies, such as SAT/SMT, QBF, Planning and Scheduling, advancements in ASP solving are assessed in competition events. In this paper, we report about the design and results of the Sixth ASP Competition, which was jointly organized by the University of Calabria (Italy), Aalto University (Finland), and the University of Genoa (Italy), in affiliation with the 13th International Conference on Logic Programming and Non-Monotonic Reasoning. This edition maintained some of the design decisions introduced in 2014, e.g., the conception of sub-tracks, the scoring scheme,and the adherence to a fixed modeling language in order to push the adoption of the ASP-Core-2 standard. On the other hand, it featured also some novelties, like a benchmark selection stage classifying instances according to their empirical hardness, and a “Marathon” track where the topperforming systems are given more time for solving hard benchmarks.
Since 2004, increases in computational power described by Moore's law have substantially been realized in the form of additional cores rather than through faster clock speeds. To make effective use of modern hardware when solving hard computational problems, it is therefore necessary to employ parallel solution strategies. In this work, we demonstrate how effective parallel solvers for propositional satisfiability (SAT), one of the most widely studied NP-complete problems, can be produced automatically from any existing sequential, highly parametric SAT solver. Our Automatic Construction of Parallel Portfolios (ACPP) approach uses an automatic algorithm configuration procedure to identify a set of configurations that perform well when executed in parallel. Applied to two prominent SAT solvers, Lingeling and clasp, our ACPP procedure identified 8-core solvers that significantly outperformed their sequential counterparts on a diverse set of instances from the application and hard combinatorial category of the 2012 SAT Challenge. We further extended our ACPP approach to produce parallel portfolio solvers consisting of several different solvers by combining their configuration spaces. Applied to the component solvers of the 2012 SAT Challenge gold medal winning SAT Solver pfolioUZK, our ACPP procedures produced a significantly better-performing parallel SAT solver.
THIS INSTALLMENT OF Research for Practice provides curated reading guides to technology for underserved communities and to new developments in personal fabrication. First, Tawanna Dillahunt describes design considerations and technology for underserved and impoverished communities. Designing for the more than 1.6 billion impoverished individuals worldwide requires special consideration of community needs, constraints, and context. Her selections span protocols for poor-quality communication networks, community-driven content generation, and resource and public service discovery. Second, Stefanie Mueller and Patrick Baudisch provide an overview of recent advances in personal fabrication (for example, 3D printers).
Answer Set Programming faces an increasing popularity for problem solving in various domains. While its modeling language allows us to express many complex problems in an easy way, its solving technology enables their effective resolution. In what follows, we detail some of the key factors of its success. Answer Set Programming [ASP; Brewka et al. Commun ACM 54(12):92–103, (2011)] is seeing a rapid proliferation in academia and industry due to its easy and flexible way to model and solve knowledge-intense combinatorial (optimization) problems. To this end, ASP offers a high-level modeling language paired with high-performance solving technology. As a result, ASP systems provide out-off-the-box, general-purpose search engines that allow for enumerating (optimal) solutions. They are represented as answer sets, each being a set of atoms representing a solution. The declarative approach of ASP allows a user to concentrate on a problem’s specification rather than the computational means to solve it. This makes ASP a prime candidate for rapid prototyping and an attractive tool for teaching key AI techniques since complex problems can be expressed in a succinct and elaboration tolerant way. This is eased by the tuning of ASP’s modeling language to knowledge representation and reasoning (KRR). The resulting impact is nicely reflected by a growing range of successful applications of ASP [Erdem et al. AI Mag 37(3):53–68, 2016; Falkner et al. Industrial applications of answer set programming. K++nstliche Intelligenz (2018)]
Nowadays, business processes are increasingly supported by IT services that produce massive amounts of event data during the execution of a process. These event data can be used to analyze the process using process mining techniques to discover the real process, measure conformance to a given process model, or to enhance existing models with performance information. Mapping the produced events to activities of a given process model is essential for conformance checking, annotation and understanding of process mining results. In order to accomplish this mapping with low manual effort, we developed a semi-automatic approach that maps events to activities using insights from behavioral analysis and label analysis. The approach extracts Declare constraints from both the log and the model to build matching constraints to efficiently reduce the number of possible mappings. These mappings are further reduced using techniques from natural language processing, which allow for a matching based on labels and external knowledge sources. The evaluation with synthetic and real-life data demonstrates the effectiveness of the approach and its robustness toward non-conforming execution logs.
In computer science, computer systems are both, objects of investigation and tools that enable creative learning and design. Tools for learning have a long tradition in computer science education. Already in the late 1960s, Papert developed a concept which had an immense impact on the development of informal education in the following years: his theory of constructionism understands learning as a creative process of knowledge construction that is most effective when learners create something purposeful that they can try out, show around, discuss, analyse and receive praise for. By now, there are numerous learning and programming environments that are based on the constructionist ideas. Modern tools offer opportunities for students to learn in motivating ways and gain impressive results in programming games, animations, implementing 3D models or developing interactive objects. This article gives an overview of computer science education research related to tools and media to be used in educational settings. We analyse different types of tools with a special focus on the categorization and development of tools for student adequate physical computing activities in the classroom. Research around the development and evaluation of tools and learning resources in the domain of physical computing is illustrated with the example of "My Interactive Garden", a constructionist learning and programming environment. It is explained how the results from empirical studies are integrated in the continuous development of the learning material.
Automated storage and retrieval systems are principal components of modern production and warehouse facilities. In particular, automated guided vehicles nowadays substitute human-operated pallet trucks in transporting production materials between storage locations and assembly stations. While low-level control systems take care of navigating such driverless vehicles along programmed routes and avoid collisions even under unforeseen circumstances, in the common case of multiple vehicles sharing the same operation area, the problem remains how to set up routes such that a collection of transport tasks is accomplished most effectively. We address this prevalent problem in the context of car assembly at Mercedes-Benz Ludwigsfelde GmbH, a large-scale producer of commercial vehicles, where routes for automated guided vehicles used in the production process have traditionally been hand-coded by human engineers. Such adhoc methods may suffice as long as a running production process remains in place, while any change in the factory layout or production targets necessitates tedious manual reconfiguration, not to mention the missing portability between different production plants. Unlike this, we propose a declarative approach based on Answer Set Programming to optimize the routes taken by automated guided vehicles for accomplishing transport tasks. The advantages include a transparent and executable problem formalization, provable optimality of routes relative to objective criteria, as well as elaboration tolerance towards particular factory layouts and production targets. Moreover, we demonstrate that our approach is efficient enough to deal with the transport tasks evolving in realistic production processes at the car factory of Mercedes-Benz Ludwigsfelde GmbH.
The Potsdam answer set solving collection, or Potassco for short, bundles various tools implementing and/or applying answer set programming. The article at hand succeeds an earlier description of the Potassco project published in Gebser et al. (AI Commun 24(2):107-124, 2011). Hence, we concentrate in what follows on the major features of the most recent, fifth generation of the ASP system clingo and highlight some recent resulting application systems.
The aim of our project design space exploration with answer set programming is to develop a general framework based on Answer Set Programming (ASP) that finds valid solutions to the system design problem and simultaneously performs Design Space Exploration (DSE) to find the most favorable alternatives. We leverage recent developments in ASP solving that allow for tight integration of background theories to create a holistic framework for effective DSE.
Novel two-dimensional tactile displays enable blind users to not only get access to the textual but also to the graphical content of a graphical user interface. Due to the higher amount of information that can be presented in parallel, orientation and exploration can be more complex. In this paper we present the HyperBraille system, which consists of a pin-matrix device as well as a graphical screen reader providing the user with appropriate presentation and interaction possibilities. To allow for a detailed analysis of bimanual interaction strategies on a pin-matrix device, we conducted two user studies with a total of 12 blind people. The task was to fill in .pdf forms on the pin-matrix device by using different input methods, namely gestures, built-in hardware buttons as well as a conventional PC keyboard. The forms were presented in a semigraphic view type that not only contains Braille but also tactile widgets in a spatial arrangement. While completion time and error rate partly depended on the chosen input method, the usage of special reading strategies seemed to be independent of it. A direct comparison of the system and a conventional assistive technology (screen reader with single-line Braille device) showed that interaction on the pin-matrix device can be very efficient if the user is trained. The two-dimensional output can improve access to .pdf forms with insufficient accessibility as the mapping of input controls and the corresponding labels can be supported by a spatial presentation.
Parsing of argumentative structures has become a very active line of research in recent years. Like discourse parsing or any other natural language task that requires prediction of linguistic structures, most approaches choose to learn a local model and then perform global decoding over the local probability distributions, often imposing constraints that are specific to the task at hand. Specifically for argumentation parsing, two decoding approaches have been recently proposed: Minimum Spanning Trees (MST) and Integer Linear Programming (ILP), following similar trends in discourse parsing. In contrast to discourse parsing though, where trees are not always used as underlying annotation schemes, argumentation structures so far have always been represented with trees. Using the ‘argumentative microtext corpus’ [in: Argumentation and Reasoned Action: Proceedings of the 1st European Conference on Argumentation, Lisbon 2015 / Vol. 2, College Publications, London, 2016, pp. 801–815] as underlying data and replicating three different decoding mechanisms, in this paper we propose a novel ILP decoder and an extension to our earlier MST work, and then thoroughly compare the approaches. The result is that our new decoder outperforms related work in important respects, and that in general, ILP and MST yield very similar performance.
teaspoon
(2018)
Answer Set Programming (ASP) is an approach to declarative problem solving, combining a rich yet simple modeling language with high performance solving capacities. We here develop an ASP-based approach to curriculum-based course timetabling (CB-CTT), one of the most widely studied course timetabling problems. The resulting teaspoon system reads a CB-CTT instance of a standard input format and converts it into a set of ASP facts. In turn, these facts are combined with a first-order encoding for CB-CTT solving, which can subsequently be solved by any off-the-shelf ASP systems. We establish the competitiveness of our approach by empirically contrasting it to the best known bounds obtained so far via dedicated implementations. Furthermore, we extend the teaspoon system to multi-objective course timetabling and consider minimal perturbation problems.
We introduce a type and effect system, for an imperative object calculus, which infers sharing possibly introduced by the evaluation of an expression, represented as an equivalence relation among its free variables. This direct representation of sharing effects at the syntactic level allows us to express in a natural way, and to generalize, widely-used notions in literature, notably uniqueness and borrowing. Moreover, the calculus is pure in the sense that reduction is defined on language terms only, since they directly encode store. The advantage of this non-standard execution model with respect to a behaviorally equivalent standard model using a global auxiliary structure is that reachability relations among references are partly encoded by scoping. (C) 2018 Elsevier B.V. All rights reserved.
Digitale Medien enthalten bislang vor allem Inhalte in verschiedenen Darstellungsformen. Dies allein erzeugt jedoch nur einen geringen Mehrwert zu klassischen Lernressourcen, da die Kriterien der Interaktivität und Adaptivität nicht mit einbezogen werden. Dies scheitert jedoch oft an dem damit verbundenen Erstellungsaufwand. Der folgende Beitrag zeigt, wie durch die automatische Erzeugung von Aufgaben ein hochwertiger Wissenserwerb mit digitalen Medien ermöglicht wird. Ferner werden Vor- und Nachteile der automatischen Erstellung von Aufgaben erörtert.
Die Veröffentlichung eines global frei verfügbaren Onlinekurses abseits der großen MOOC Plattformen bringt spezielle Herausforderungen mit sich. Neben technischen Herausforderungen sind eine effiziente Wissensvermittlung und die Erhaltung der Lernmotivation zentral. Der folgende Beitrag stellt Techniken zur Steigerung der Lerneffizienz und -motivation anhand des ARCS Modells vor. Er zeigt auf, wie die verschiedenen Techniken in der Entwicklung des Onlinekurses „Designing Sustainable Food Systems“ umgesetzt wurden und inwieweit sie erfolgreich waren.
Das Training sozioemotionaler Kompetenzen ist gerade für Menschen mit Autismus nützlich. Ein solches Training kann mithilfe einer spielbasierten Anwendung effektiv gestaltet werden. Zwei Minispiele, Mimikry und Emo-Mahjong, wurden realisiert und hinsichtlich User Experience evaluiert. Die jeweiligen Konzepte und die Evaluationsergebnisse sollen hier vorgestellt werden.
Einsatz einer mobilen Lern-App - Ein Werkzeug zur Verbesserung von klinisch-praktischem Unterricht
(2018)
Der Unterricht am Krankenbett ist im Medizinstudium eine wertvolle Möglichkeit klinisch-praktische Fertigkeiten zu erlernen. Eine optimale Vorbereitung der Studierenden ist dabei Voraussetzung. Eine mobile Lern-App wurde entwickelt, die den Studierenden, neben Lernzielen, Kursinhalte und Anleitungen zu Untersuchungstechniken bietet, um die Vorbereitung auf einen klinisch-praktischen Kurs zu fördern und Kurzinformationen auch während des Kurses zur Verfügung zu stellen. 175 Studierende hatten die Möglichkeit die App parallel zu einem klinischen Untersuchungs-Kurs im Semester zu nutzen. Im Anschluss beantworteten die Studierenden einen Fragebogen zur Nützlichkeit und Vielseitigkeit der App und zur Zufriedenheit mit der App unter Verwendung eine 5-Punkt-Likert-Skala und zwei offenen Fragen. In diesem Beitrag wird das Kurskonzept zusammen mit der Lern-App, die Ergebnisse aus dem Fragebogen und unsere Schlussfolgerungen daraus vorgestellt. Studierende bewerteten die App grundsätzlich als hilfreich. Sie sollte dabei gründlich eingeführt werden. Patienten sollten über die Nutzung von Smartphones im Studentenunterricht zu Lernzwecken informiert werden.
Der vorliegende Beitrag berichtet auf der Grundlage von Erfahrungen mit dem Audience Response System (ARS) „Auditorium Mobile Classroom Service“ von Erfolgsfaktoren für den Einsatz in der universitären Lehre. Dabei werden sowohl die technischen Rahmenbedingungen und Herausforderungen der Anwendungen berücksichtigt, als auch die unterschiedlichen didaktischen Konzepte und Ziele der beteiligten Akteure (Studierende, Lehrende und Institution). Ziel ist es, Einflussfaktoren für den erfolgreichen Einsatz sowohl für die Praxis als auch die wissenschaftliche Untersuchung und Weiterentwicklung der Systeme zu benennen und ein heuristisches Framework für Chancen und Herausforderungen beim Einsatz von ARS anzubieten.
Der Beitrag skizziert ein Modell, das die Entwicklung digitaler Kompetenzen im Lehramtsstudium fördern soll. Zwar wird das Kompetenzmodell aus der Deutschdidaktik heraus entwickelt, nimmt aber auch fachübergreifende Anforderungen in den Bereichen Informationskompetenz, medientechnischer Kompetenzen, Fähigkeiten der Medienanalyse und -reflexion sowie Sprachhandlungskompetenz in den Blick. Damit wird das Ziel verfolgt, die besonderen Anforderungen angehender Lehrkräfte als Mediator*innen digitaler Kompetenzen darzustellen. Das beschriebene Modell dieser Vermittlungskompetenz dient der Verankerung digitaler Lehr-Lernkonzepte als wesentlicher Bestandteil der modernen Lehrer*innenbildung.
Ob Online-Kurse, videobasierte Lehrangebote, mobile Applikationen, eigenentwickelte oder kommerzielle Web 2.0-Anwendungen, die Fülle digitaler Unterstützungsangebote ist kaum zu überblicken. Dabei bieten mobile Endgeräte, Web-Anwendungen und Apps Chancen Lehre, Studium und Forschung maßgeblich neu zu gestalten. Im Beitrag wird ein Beschreibungsrahmen für die mediendidaktische Ausgestaltung von Lehr-, Lern- und Forschungsarrangements vorgestellt, der die technischen Gesichtspunkte hervorhebt. Anschließend werden unterschiedliche Nutzungsszenarien unter Einbeziehung digitaler Medien skizziert. Diese werden als Ausgangspunkt genommen um das Konzept einer Systemarchitektur vorzustellen, die es zum einen ermöglicht beliebige Applikationen automatisiert bereit zu stellen und zum anderen die anfallenden Nutzendendaten plattformübergreifend zu aggregieren und für eine Ausgestaltung virtueller Lehr- und Lernräumen zu nutzen.
In this paper, we consider the computational power of a new variant of networks of splicing processors in which each processor as well as the data navigating throughout the network are now considered to be polarized. While the polarization of every processor is predefined (negative, neutral, positive), the polarization of data is dynamically computed by means of a valuation mapping. Consequently, the protocol of communication is naturally defined by means of this polarization. We show that networks of polarized splicing processors (NPSP) of size 2 are computationally complete, which immediately settles the question of designing computationally complete NPSPs of minimal size. With two more nodes we can simulate every nondeterministic Turing machine without increasing the time complexity. Particularly, we prove that NPSP of size 4 can accept all languages in NP in polynomial time. Furthermore, another computational model that is universal, namely the 2-tag system, can be simulated by NPSP of size 3 preserving the time complexity. All these results can be obtained with NPSPs with valuations in the set as well. We finally show that Turing machines can simulate a variant of NPSPs and discuss the time complexity of this simulation.
Berufsbegleitende Studiengänge stehen vor besonderen Schwierigkeiten, für die der Einsatz von Blended Learning-Szenarien sinnvoll sein kann. Welche speziellen Herausforderungen sich dabei ergeben und welche Lösungsansätze dagegen steuern, betrachtet der folgende Artikel anhand eines Praxisberichts aus dem Studiengang M. P. A. Wissenschaftsmanagement an der Universität Speyer.
The soft error rate (SER) due to heavy-ion irradiation of a clock tree is investigated in this paper. A method for clock tree SER prediction is developed, which employs a dedicated soft error analysis tool to characterize the single-event transient (SET) sensitivities of clock inverters and other commercial tools to calculate the SER through fault-injection simulations. A test circuit including a flip-flop chain and clock tree in a 65 nm CMOS technology is developed through the automatic ASIC design flow. This circuit is analyzed with the developed method to calculate its clock tree SER. In addition, this circuit is implemented in a 65 nm test chip and irradiated by heavy ions to measure its SER resulting from the SETs in the clock tree. The experimental and calculation results of this case study present good correlation, which verifies the effectiveness of the developed method.
Detection of malware-infected computers and detection of malicious web domains based on their encrypted HTTPS traffic are challenging problems, because only addresses, timestamps, and data volumes are observable. The detection problems are coupled, because infected clients tend to interact with malicious domains. Traffic data can be collected at a large scale, and antivirus tools can be used to identify infected clients in retrospect. Domains, by contrast, have to be labeled individually after forensic analysis. We explore transfer learning based on sluice networks; this allows the detection models to bootstrap each other. In a large-scale experimental study, we find that the model outperforms known reference models and detects previously unknown malware, previously unknown malware families, and previously unknown malicious domains.
Answer Set Programming (ASP) has become a popular and widespread paradigm for practical Knowledge Representation thanks to its expressiveness and the available enhancements of its input language. One of such enhancements is the use of aggregates, for which different semantic proposals have been made. In this paper, we show that any ASP aggregate interpreted under Gelfond and Zhang's (GZ) semantics can be replaced (under strong equivalence) by a propositional formula. Restricted to the original GZ syntax, the resulting formula is reducible to a disjunction of conjunctions of literals but the formulation is still applicable even when the syntax is extended to allow for arbitrary formulas (including nested aggregates) in the condition. Once GZ-aggregates are represented as formulas, we establish a formal comparison (in terms of the logic of Here-and-There) to Ferraris' (F) aggregates, which are defined by a different formula translation involving nested implications. In particular, we prove that if we replace an F-aggregate by a GZ-aggregate in a rule head, we do not lose answer sets (although more can be gained). This extends the previously known result that the opposite happens in rule bodies, i.e., replacing a GZ-aggregate by an F-aggregate in the body may yield more answer sets. Finally, we characterize a class of aggregates for which GZ- and F-semantics coincide.
The Surface Water and Ocean Topography (SWOT) mission is a next generation satellite mission expected to provide a 2 km-resolution observation of the sea surface height (SSH) on a two-dimensional swath. Processing SWOT data will be challenging because of the large amount of data, the mismatch between a high spatial resolution and a low temporal resolution, and the observation errors. The present paper focuses on the reduction of the spatially structured errors of SWOT SSH data. It investigates a new error reduction method and assesses its performance in an observing system simulation experiment. The proposed error-reduction method first projects the SWOT SSH onto a subspace spanned by the SWOT spatially structured errors. This projection is removed from the SWOT SSH to obtain a detrended SSH. The detrended SSH is then processed within an ensemble data assimilation analysis to retrieve a full SSH field. In the latter step, the detrending is applied to both the SWOT data and an ensemble of model-simulated SSH fields. Numerical experiments are performed with synthetic SWOT observations and an ensemble from a North Atlantic, 1/60 degrees simulation of the ocean circulation (NATL60). The data assimilation analysis is carried out with an ensemble Kalman filter. The results are assessed with root mean square errors, power spectrum density, and spatial coherence. They show that a significant part of the large scale SWOT errors is reduced. The filter analysis also reduces the small scale errors and allows for an accurate recovery of the energy of the signal down to 25 km scales. In addition, using the SWOT nadir data to adjust the SSH detrending further reduces the errors.
plasp 3
(2019)
We describe the new version of the Planning Domain Definition Language (PDDL)-to-Answer Set Programming (ASP) translator plasp. First, it widens the range of accepted PDDL features. Second, it contains novel planning encodings, some inspired by Satisfiability Testing (SAT) planning and others exploiting ASP features such as well-foundedness. All of them are designed for handling multivalued fluents in order to capture both PDDL as well as SAS planning formats. Third, enabled by multishot ASP solving, it offers advanced planning algorithms also borrowed from SAT planning. As a result, plasp provides us with an ASP-based framework for studying a variety of planning techniques in a uniform setting. Finally, we demonstrate in an empirical analysis that these techniques have a significant impact on the performance of ASP planning.
Multi-sided platforms (MSP) strongly affect markets and play a crucial part within the digital and networked economy. Although empirical evidence indicates their occurrence in many industries, research has not investigated the game-changing impact of MSP on traditional markets to a sufficient extent. More specifically, we have little knowledge of how MSP affect value creation and customer interaction in entire markets, exploiting the potential of digital technologies to offer new value propositions. Our paper addresses this research gap and provides an initial systematic approach to analyze the impact of MSP on the insurance industry. For this purpose, we analyze the state of the art in research and practice in order to develop a reference model of the value network for the insurance industry. On this basis, we conduct a case-study analysis to discover and analyze roles which are occupied or even newly created by MSP. As a final step, we categorize MSP with regard to their relation to traditional insurance companies, resulting in a classification scheme with four MSP standard types: Competition, Coordination, Cooperation, Collaboration.
In this work we tackle the problem of checking strong equivalence of logic programs that may contain local auxiliary atoms, to be removed from their stable models and to be forbidden in any external context. We call this property projective strong equivalence (PSE). It has been recently proved that not any logic program containing auxiliary atoms can be reformulated, under PSE, as another logic program or formula without them – this is known as strongly persistent forgetting. In this paper, we introduce a conservative extension of Equilibrium Logic and its monotonic basis, the logic of Here-and-There, in which we deal with a new connective ‘|’ we call fork. We provide a semantic characterisation of PSE for forks and use it to show that, in this extension, it is always possible to forget auxiliary atoms under strong persistence. We further define when the obtained fork is representable as a regular formula.
In recent years, named entity linking (NEL) tools were primarily developed in terms of a general approach, whereas today numerous tools are focusing on specific domains such as e.g. the mapping of persons and organizations only, or the annotation of locations or events in microposts. However, the available benchmark datasets necessary for the evaluation of NEL tools do not reflect this focalizing trend. We have analyzed the evaluation process applied in the NEL benchmarking framework GERBIL [in: Proceedings of the 24th International Conference on World Wide Web (WWW’15), International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, Switzerland, 2015, pp. 1133–1143, Semantic Web 9(5) (2018), 605–625] and all its benchmark datasets. Based on these insights we have extended the GERBIL framework to enable a more fine grained evaluation and in depth analysis of the available benchmark datasets with respect to different emphases. This paper presents the implementation of an adaptive filter for arbitrary entities and customized benchmark creation as well as the automated determination of typical NEL benchmark dataset properties, such as the extent of content-related ambiguity and diversity. These properties are integrated on different levels, which also enables to tailor customized new datasets out of the existing ones by remixing documents based on desired emphases. Besides a new system library to enrich provided NIF [in: International Semantic Web Conference (ISWC’13), Lecture Notes in Computer Science, Vol. 8219, Springer, Berlin, Heidelberg, 2013, pp. 98–113] datasets with statistical information, best practices for dataset remixing are presented, and an in depth analysis of the performance of entity linking systems on special focus datasets is presented.
Metabolic networks play a crucial role in biology since they capture all chemical reactions in an organism. While there are networks of high quality for many model organisms, networks for less studied organisms are often of poor quality and suffer from incompleteness. To this end, we introduced in previous work an answer set programming (ASP)-based approach to metabolic network completion. Although this qualitative approach allows for restoring moderately degraded networks, it fails to restore highly degraded ones. This is because it ignores quantitative constraints capturing reaction rates. To address this problem, we propose a hybrid approach to metabolic network completion that integrates our qualitative ASP approach with quantitative means for capturing reaction rates. We begin by formally reconciling existing stoichiometric and topological approaches to network completion in a unified formalism. With it, we develop a hybrid ASP encoding and rely upon the theory reasoning capacities of the ASP system dingo for solving the resulting logic program with linear constraints over reals. We empirically evaluate our approach by means of the metabolic network of Escherichia coli. Our analysis shows that our novel approach yields greatly superior results than obtainable from purely qualitative or quantitative approaches.
PLATON
(2019)
Lesson planning is both an important and demanding task—especially as part of teacher training. This paper presents the requirements for a lesson planning system and evaluates existing systems regarding these requirements. One major drawback of existing software tools is that most are limited to a text- or form-based representation of the lesson designs. In this article, a new approach with a graphical, time-based representation with (automatic) analyses methods is proposed and the system architecture and domain model are described in detail. The approach is implemented in an interactive, web-based prototype called PLATON, which additionally supports the management of lessons in units as well as the modelling of teacher and student-generated resources. The prototype was evaluated in a study with 61 prospective teachers (bachelor’s and master’s preservice teachers as well as teacher trainees in post-university teacher training) in Berlin, Germany, with a focus on usability. The results show that this approach proofed usable for lesson planning and offers positive effects for the perception of time and self-reflection.
Answer Set Programming (ASP) is a successful rule-based formalism for modeling and solving knowledge-intense combinatorial (optimization) problems. Despite its success in both academic and industry, open challenges like automatic source code optimization, and software engineering remains. This is because a problem encoded into an ASP might not have the desired solving performance compared to an equivalent representation. Motivated by these two challenges, this paper has three main contributions. First, we propose a developing process towards a methodology to implement ASP programs, being faithful to existing methods. Second, we present ASP encodings that serve as the basis from the developing process. Third, we demonstrate the use of ASP to reverse the standard solving process. That is, knowing answer sets in advance, and desired strong equivalent properties, “we” exhaustively reconstruct ASP programs if they exist. This paper was originally motivated by the search of propositional formulas (if they exist) that represent the semantics of a new aggregate operator. Particularly, a parity aggregate. This aggregate comes as an improvement from the already existing parity (xor) constraints from xorro, where lacks expressiveness, even though these constraints fit perfectly for reasoning modes like sampling or model counting. To this end, this extended version covers the fundaments from parity constraints as well as the xorro system. Hence, we delve a little more in the examples and the proposed methodology over parity constraints. Finally, we discuss our results by showing the only representation available, that satisfies different properties from the classical logic xor operator, which is also consistent with the semantics of parity constraints from xorro.
Research publications and data nowadays should be publicly available on the internet and, theoretically, usable for everyone to develop further research, products, or services. The long-term accessibility of research data is, therefore, fundamental in the economy of the research production process. However, the availability of data is not sufficient by itself, but also their quality must be verifiable. Measures to ensure reuse and reproducibility need to include the entire research life cycle, from the experimental design to the generation of data, quality control, statistical analysis, interpretation, and validation of the results. Hence, high-quality records, particularly for providing a string of documents for the verifiable origin of data, are essential elements that can act as a certificate for potential users (customers). These records also improve the traceability and transparency of data and processes, therefore, improving the reliability of results. Standards for data acquisition, analysis, and documentation have been fostered in the last decade driven by grassroot initiatives of researchers and organizations such as the Research Data Alliance (RDA). Nevertheless, what is still largely missing in the life science academic research are agreed procedures for complex routine research workflows. Here, well-crafted documentation like standard operating procedures (SOPs) offer clear direction and instructions specifically designed to avoid deviations as an absolute necessity for reproducibility. Therefore, this paper provides a standardized workflow that explains step by step how to write an SOP to be used as a starting point for appropriate research documentation.
The notion of coherence relations is quite widely accepted in general, but concrete proposals differ considerably on the questions of how they should be motivated, which relations are to be assumed, and how they should be defined. This paper takes a "bottom-up" perspective by assessing the contribution made by linguistic signals (connectives), using insights from the relevant literature as well as verification by practical text annotation. We work primarily with the German language here and focus on the realm of contrast. Thus, we suggest a new inventory of contrastive connective functions and discuss their relationship to contrastive coherence relations that have been proposed in earlier work.
Many Android applications embed webpages via WebView components and execute JavaScript code within Android. Hybrid applications leverage dedicated APIs to load a resource and render it in a WebView. Furthermore, Android objects can be shared with the JavaScript world. However, bridging the interfaces of the Android and JavaScript world might also incur severe security threats: Potentially untrusted webpages and their JavaScript might interfere with the Android environment and its access to native features.
No general analysis is currently available to assess the implications of such hybrid apps bridging the two worlds. To understand the semantics and effects of hybrid apps, we perform a large-scale study on the usage of the hybridization APIs in the wild. We analyze and categorize the parameters to hybridization APIs for 7,500 randomly selected and the 196 most popular applications from the Google Playstore as well as 1000 malware samples. Our results advance the general understanding of hybrid applications, as well as implications for potential program analyses, and the current security situation: We discovered thousands of flows of sensitive data from Android to JavaScript, the vast majority of which could flow to potentially untrustworthy code. Our analysis identified numerous web pages embedding vulnerabilities, which we exemplarily exploited. Additionally, we discovered a multitude of applications in which potentially untrusted JavaScript code may interfere with (trusted) Android objects, both in benign and malign applications.
M-rate 0L systems are interactionless Lindenmayer systems together with a function assigning to every string a set of multisets of productions that may be applied simultaneously to the string. Some questions that have been left open in the forerunner papers are examined, and the computational power of deterministic M-rate 0L systems is investigated, where also tabled and extended variants are taken into consideration.
We study the derivational complexity of context-free and context-sensitive grammars by counting the maximal number of non-regular and non-context-free rules used in a derivation, respectively. The degree of non-regularity/non-context-freeness of a language is the minimum degree of non-regularity/non-context-freeness of context-free/context-sensitive grammars generating it. A language has finite degree of non-regularity iff it is regular. We give a condition for deciding whether the degree of non-regularity of a given unambiguous context-free grammar is finite. The problem becomes undecidable for arbitrary linear context-free grammars. The degree of non-regularity of unambiguous context-free grammars generating non-regular languages as well as that of grammars generating deterministic context-free languages that are not regular is of order Omega(n). Context-free non-regular languages of sublinear degree of non-regularity are presented. A language has finite degree of non-context-freeness if it is context-free. Context-sensitive grammars with a quadratic degree of non-context-freeness are more powerful than those of a linear degree.
Many knowledge representation tasks involve trees or similar structures as abstract datatypes. However, devising compact and efficient declarative representations of such structural properties is non-obvious and can be challenging indeed. In this article, we take a number of acyclicity properties into consideration and investigate various logic-based approaches to encode them. We use answer set programming as the primary representation language but also consider mappings to related formalisms, such as propositional logic, difference logic and linear programming. We study the compactness of encodings and the resulting computational performance on benchmarks involving acyclic or tree structures.