Institut für Informatik und Computational Science
Refine
Year of publication
- 2023 (2)
- 2022 (18)
- 2021 (16)
- 2020 (22)
- 2019 (10)
- 2018 (21)
- 2017 (4)
- 2016 (9)
- 2015 (63)
- 2014 (29)
- 2013 (37)
- 2012 (23)
- 2011 (27)
- 2010 (32)
- 2009 (26)
- 2008 (1)
- 2007 (24)
- 2006 (38)
- 2005 (29)
- 2004 (27)
- 2003 (18)
- 2002 (15)
- 2001 (20)
- 2000 (31)
- 1999 (29)
- 1998 (30)
- 1997 (22)
- 1996 (25)
- 1995 (15)
- 1994 (17)
- 1993 (1)
Document Type
- Article (682) (remove)
Keywords
- Didaktik (13)
- Ausbildung (12)
- Hochschuldidaktik (12)
- Informatik (12)
- Answer set programming (10)
- answer set programming (8)
- Answer Set Programming (6)
- Computer Science Education (4)
- E-Learning (4)
- Competence Measurement (3)
- Machine learning (3)
- Secondary Education (3)
- Theory (3)
- education (3)
- formal languages (3)
- monitoring (3)
- Analytical models (2)
- Automata systems (2)
- Big Data (2)
- Competence Modelling (2)
- Computational thinking (2)
- Digitale Medien (2)
- E-learning (2)
- Equilibrium logic (2)
- Event mapping (2)
- Fault tolerance (2)
- IT-Infrastruktur (2)
- Informatics Education (2)
- Informatics Modelling (2)
- Informatics System Application (2)
- Informatics System Comprehension (2)
- Lindenmayer systems (2)
- Non-monotonic reasoning (2)
- Parameterized complexity (2)
- Process mining (2)
- ResNet (2)
- Strategie (2)
- bioinformatics (2)
- computational thinking (2)
- computer science education (2)
- cooperating systems (2)
- e-learning (2)
- higher education (2)
- knowledge representation and nonmonotonic reasoning (2)
- online learning (2)
- radhard design (2)
- relevance (2)
- security (2)
- soft errors (2)
- tools (2)
- (FPGA) (1)
- (SET) count rate (1)
- 2-tag system (1)
- 21st century skills, (1)
- 3D modeling (1)
- 3D visualization (1)
- ABRACADABRA (1)
- AODV (1)
- ARCS Modell (1)
- ASIC (1)
- Absorbed dose (1)
- Abstraction (1)
- Access control (1)
- Achievement (1)
- Active evaluation (1)
- Activity Theory (1)
- Activity-orientated Learning (1)
- Ad hoc routing (1)
- Adaptivity (1)
- Adaptivität (1)
- Advanced Video Codec (AVC) (1)
- Aggregates (1)
- Algorithm configuration (1)
- Algorithm portfolios (1)
- Algorithms (1)
- Android hybrid apps (1)
- Animal building (1)
- Anti-cancer drugs (1)
- Arduino (1)
- Argumentation structure (1)
- Assessment (1)
- Audience Response Systeme (1)
- Augmentation (1)
- Augmented and virtual reality (1)
- Austria (1)
- Autismus (1)
- Automated Theorem Proving (1)
- Automated parallelization (1)
- Automatic Item Generation (1)
- Automatically controlled windows (1)
- Automatisches Beweisen (1)
- Backdoors (1)
- Batch processing (1)
- Bean (1)
- Benchmark testing; (1)
- Berliner Modell (1)
- Blended Learning (1)
- Blind users (1)
- Bloom’s Taxonomy (1)
- Boolean logic models (1)
- Business process intelligence (1)
- CP-Logic (1)
- CS concepts (1)
- Campus (1)
- Capability approach (1)
- Challenges (1)
- Circuit faults (1)
- Clause Learning (1)
- Clock tree (1)
- Cloud (1)
- Cluster Computing (1)
- Cluster computing (1)
- Code generation (1)
- Cognitive Skills (1)
- Coherent phonons (1)
- Combinatorial multi-objective optimization (1)
- Competences (1)
- Competencies (1)
- Complex optimization (1)
- Complexity (1)
- Computational Thinking (1)
- Computational complexity (1)
- Computational grid (1)
- Computer Science (1)
- Computer Science in Context (1)
- Computer security (1)
- Computergestützes Training (1)
- Computing (1)
- Computing with DNA (1)
- Conformant Planning (1)
- Conrad Hal Waddington (1)
- Constraint satisfaction (1)
- Contest (1)
- Context awareness (1)
- Contextualisation (1)
- Contextualized learning (1)
- Continuous Testing (1)
- Continuous Versioning (1)
- Contradictions (1)
- Convolution (1)
- Course timetabling (1)
- Curriculum (1)
- Curriculum Development (1)
- Customer ownership (1)
- D-galactosamine (1)
- DMR (1)
- DNA hairpin formation (1)
- DPLL (1)
- DRMAA (1)
- DRMS (1)
- Data Analysis (1)
- Data Management (1)
- Data Privacy (1)
- Data federation (1)
- Database (1)
- Databases (1)
- Deal of the Day (1)
- Debugging (1)
- Decidability (1)
- Declare (1)
- Deep learning (1)
- Defining characteristics of physical computing (1)
- Denotational semantics (1)
- Design (1)
- Design for testability (DFT) (1)
- Digital Competence (1)
- Digital Education (1)
- Digital Game Based Learning (1)
- Digital Revolution (1)
- Digital image analysis (1)
- Digitalisierung (1)
- Digitalization (1)
- Dose rate (1)
- Double cell upsets (DCUs) (1)
- Dynamic assessment (1)
- Dynamical X-ray theory (1)
- E-Assessment (1)
- E-Klausuren (1)
- E-Portfolio (1)
- E-teaching (1)
- EDC (1)
- EEG (1)
- Early Literacy (1)
- Educational Standards (1)
- Educational game (1)
- Educational software (1)
- Educational timetabling (1)
- Elektronisches Prüfen (1)
- Embedded Systems (1)
- Encoding (1)
- Engines (1)
- Entity Linking (1)
- Epigenetic landscape (1)
- Epistemic Logic Programs (1)
- Euclid’s algorithm (1)
- Evaluation (1)
- Evolution (1)
- Experimentation (1)
- Explicit negation (1)
- Explore-first Programming (1)
- Extensibility (1)
- Extreme Model-Driven Development (1)
- FEDC (1)
- FPGA (1)
- Facebook (1)
- Fault Localization (1)
- Fault tolerant systems (1)
- Feature extraction (1)
- Fibonacci numbers (1)
- Field programmable gate arrays (1)
- Finite automata (1)
- Flip-flops (1)
- Forgetting (1)
- Forschungsdatenmanagement (1)
- Freshmen (1)
- Function (1)
- Fundamental Ideas (1)
- GERBIL (1)
- Gesture input (1)
- Graphensuche (1)
- Green computing (1)
- Grounded theory (1)
- H.264 (1)
- Hairpin completions (1)
- Hairpin reductions (1)
- Hardware accelerator (1)
- Heat diffusion (1)
- Hierarchically configurable mask register (1)
- Histograms (1)
- Hochschul-Cloud (1)
- Hochschullehre (1)
- Https traffic (1)
- Human Factors (1)
- Hurricane Sandy (1)
- ICT Competence (1)
- ICT competencies (1)
- ICT skills (1)
- IaaS (1)
- Identifiers (1)
- Image and video stylization (1)
- Image resolution (1)
- Imperative calculi (1)
- Improving classroom (1)
- Incoherent phonons (1)
- Incremental answer set programming (1)
- Inference (1)
- Informatics (1)
- Informatikstudium (1)
- Information federation (1)
- Information integration (1)
- Information retrieval (1)
- Information security (1)
- Innovation (1)
- Inquiry-based Learning (1)
- Insurance industry (1)
- Integrated circuit modeling (1)
- Interaktivität (1)
- Interface design (1)
- Internet of Things (1)
- Job monitoring (1)
- Job submission (1)
- Kernel (1)
- Kernelization (1)
- Key Competencies (1)
- Key input (1)
- Klausellernen (1)
- Knowledge representation (1)
- Kompetenzerwerb (1)
- L systems (1)
- LBA problem (1)
- LMS (1)
- Landmark visibility (1)
- Learners (1)
- Learning Fields (1)
- Learning ecology (1)
- Learning interfaces development (1)
- Learning with ICT (1)
- Lehrer*innenbildung (1)
- Lern-App (1)
- Lernaufgaben (1)
- Lernmotivation (1)
- Literature mining (1)
- Liver neoplasms (1)
- Load Balancing (1)
- Localization (1)
- Location awareness (1)
- Logarithm (1)
- Logic programming (1)
- Loss (1)
- Low Latency (1)
- Lower Secondary Level (1)
- Loyalty (1)
- MOOCs (1)
- MQTT (1)
- Machine Learning (1)
- Markov processes (1)
- Masking of X-values (1)
- Massive Open Online Courses (1)
- Measurement (1)
- Media in education (1)
- Meta-Programming (1)
- Metric learning (1)
- Minimal perturbation problems (1)
- Mobile App (1)
- Mobile application (1)
- Mobile devices (1)
- Mobile learning (1)
- Model checking (1)
- Modeling (1)
- Modelling (1)
- Multi-objective optimization (1)
- Multi-sided platforms (1)
- Multiple interpretation scheme (1)
- Music Technology (1)
- N-temperature model (1)
- NUI (1)
- Nash equilibrium (1)
- Natural Science Education (1)
- Natural language processing (1)
- Natural ventilation (1)
- Network (1)
- Network security (1)
- Neural networks (1)
- NoSQL (1)
- Non-Monotonic (1)
- Nonmonotonic reasoning (1)
- Norway (1)
- Novice programmers (1)
- OCCI (1)
- OSSE (1)
- OpenOLAT (1)
- Operation problem (1)
- Optimization (1)
- Parallel SAT solving (1)
- Parallel job execution time estimation (1)
- Particle detector (1)
- Partizipation (1)
- Pedagogical content knowledge (1)
- Pedagogical issues (1)
- Pedestrian navigation (1)
- Performance Evaluation (1)
- Personalization (1)
- Persönliche Lernumgebung (1)
- Pervasive computing (1)
- Pervasive game (1)
- Pervasive learning (1)
- Phantoms (1)
- Physical Science (1)
- Planar tactile display (1)
- Plant identification (1)
- Polarization (1)
- Preference Handling (1)
- Preprocessing (1)
- Problem Solving (1)
- Process model analysis (1)
- Product lifecycle management (1)
- Programming (1)
- Programming by optimization (1)
- Prototyping (1)
- RADFET (1)
- RADFETs (1)
- REST (1)
- RSA triangle (1)
- Radiation hardness (1)
- Random access memory (1)
- Ranking (1)
- Reasoning (1)
- Recommendations for CS-Curricula in Higher Education (1)
- Region of Interest (1)
- Relevanz (1)
- Reproducibility of results (1)
- Reversibility (1)
- SAT (1)
- SET pulsewidth distribution (1)
- SOA (1)
- SWOT (1)
- SaaSAbstract (1)
- Scalability (1)
- Scale-invariant feature transform (SIFT) (1)
- Scientific images (1)
- Screen reader (1)
- Seamless learning (1)
- Search problems (1)
- Security (1)
- Self-adaptive MPSoC (1)
- Semantic data (1)
- Semantic web (1)
- Semilinearity property (1)
- Sensors (1)
- Sequence embeddings (1)
- Service orientation (1)
- Sharing (1)
- Signal processing (1)
- Signaling transduction networks (1)
- Simulations (1)
- Single event effect (1)
- Single event upsets (1)
- Single-event transient (SET) (1)
- Small Private Online Courses (1)
- Social (1)
- Splicing (1)
- Splicing processor (1)
- Statistical relational learning (1)
- Stochastic relational process (1)
- Strong equivalence (1)
- Structural equation modeling (1)
- Studentenjobs (1)
- Studienabbrecher (1)
- Studiendauer (1)
- Systems biology (1)
- Systems of parallel communicating (1)
- TMR (1)
- TPACK (1)
- Tasks (1)
- Teacher perceptions (1)
- Teachers (1)
- Teaching information security (1)
- Technology proficiency (1)
- Terminology (1)
- Tests (1)
- Theorembeweisen (1)
- Theory formation (1)
- Thermoelasticity (1)
- Time series (1)
- Tomography (1)
- Tools (1)
- Tracking (1)
- Traffic data (1)
- Tree decomposition (1)
- Treewidth (1)
- Treewidth-aware reductions (1)
- Triple modular redundancy (TMR) (1)
- Tumor types (1)
- Turing machine (1)
- Type and effect systems (1)
- UAV imagery (1)
- UX (1)
- Ubiquitous learning (1)
- Ultrafast dynamics (1)
- Unary languages (1)
- Unifikation (1)
- Uniform Access Principle (1)
- Usability testing (1)
- User Experience (1)
- User submission pattern (1)
- User-centred design (1)
- VGG16 (1)
- Value network (1)
- Verification (1)
- Visual metaphor (1)
- Vocational Education (1)
- Weiterbildung (1)
- Wireless Sensor Networks (1)
- Word embeddings (1)
- X-masking (1)
- X-ray computed (1)
- X-values (1)
- Young People (1)
- abstraction (1)
- accepting grammars (1)
- action and change (1)
- activity (1)
- acute liver failure (1)
- acyclicity properties (1)
- adversarial classification (1)
- algorithm schedules (1)
- algorithms (1)
- analogical thinking (1)
- analysis (1)
- annealing (1)
- anti-cancer drugs (1)
- anxiety (1)
- approximate model counting (1)
- architecture (1)
- argument mining (1)
- arousal (1)
- artistic rendering (1)
- asynchrounous design (1)
- autism (1)
- automata (1)
- automated guided vehicle routing (1)
- automated planning (1)
- automatic feedback (1)
- behavioral abstraction (1)
- belief merging (1)
- belief revision (1)
- benchmark (1)
- bibliometric analysis (1)
- binary representation (1)
- binary search (1)
- block representation (1)
- bootstrapping (1)
- brain-computer interface (1)
- bulk built-in current sensor (1)
- bundled data (1)
- camera sensor (1)
- car assembly operations (1)
- cellular automata (1)
- circuit Faults (1)
- citation analysis (1)
- classroom language (1)
- click controller (1)
- clocks (1)
- co-citation analysis (1)
- co-occurrence analysis (1)
- code generation (1)
- cognitive modifiability (1)
- coherence relation (1)
- collaborative learning (1)
- combinatorial optimization problems (1)
- combined task and motion planning (1)
- common spatial patterns (1)
- competence (1)
- competencies (1)
- competency (1)
- competition (1)
- complexity (1)
- compliance (1)
- comprehension (1)
- computer science teachers (1)
- computer vision (1)
- concession (1)
- concurrent checking (1)
- conductive argument (1)
- connective (1)
- consistency (1)
- consistency checking (1)
- consistency measures (1)
- context-free grammar (1)
- context-sensitive (1)
- contrast (1)
- controlled vocabularies (1)
- corpus analysis (1)
- correlated errors (1)
- course timetabling (1)
- craters (1)
- crop (1)
- cs4fn (1)
- curriculum theory (1)
- decidability questions (1)
- declarative problem solving (1)
- deep learning (1)
- deep neural networks (1)
- deep residual networks (1)
- degree of non-context-freeness (1)
- degree of non-regularity (1)
- degree of non-regulation (1)
- depression (1)
- design flow (1)
- determinism (1)
- detrending (1)
- developmental systems (1)
- diagnosis (1)
- didactics (1)
- didaktisches Konzept (1)
- digitale Bildung (1)
- digitale Medien (1)
- digitally-enabled pedagogies (1)
- divide and conquer (1)
- domain-specific APIs (1)
- drug discovery (1)
- drug-sensitivity prediction (1)
- dynamic service binding (1)
- e-mentoring (1)
- eLectures (1)
- economic ripples (1)
- edge computing (1)
- education and public policy (1)
- educational programming (1)
- educational systems (1)
- educational timetabling (1)
- edutainment (1)
- embedded systems (1)
- emission factor (1)
- endothelin (1)
- endothelin-converting enzyme (1)
- ensemble kalman filter (1)
- ensemble methods (1)
- environments (1)
- error propagation (1)
- evaluation (1)
- event-related desynchronization (1)
- evolution (1)
- exponentiation (1)
- external ambiguity (1)
- extreme weather (1)
- face tracking (1)
- facial expression (1)
- fading (1)
- fault tolerance (1)
- field-programmable gate array (1)
- finite model computation (1)
- finite state sequential transducers (1)
- firmware update (1)
- formal (1)
- formal argumentation systems (1)
- fun (1)
- functions (1)
- gap-filling (1)
- geovisualization (1)
- governance (1)
- gradient boosting (1)
- grammar (1)
- graph-search (1)
- greenhouse gas (1)
- hardware accelerator (1)
- hardware architecture (1)
- high school (1)
- higher (1)
- hybrid solving (1)
- ice harboring (1)
- image classification (1)
- image processing (1)
- image recognition (1)
- imaging (1)
- impacts (1)
- incremental SVM (1)
- informal and formal learning (1)
- informal logic (1)
- informatics (1)
- informatics education (1)
- information and communication technologies (1)
- information flow control (1)
- innovation (1)
- interactive course (1)
- interactive workshop (1)
- internal ambiguity (1)
- intrusion detection (1)
- irradiation (1)
- key competences in physical computing (1)
- key competencies (1)
- kidney cancer (1)
- kinaesthetic teaching (1)
- klinisch-praktischer Unterricht (1)
- knowledge representation and reasoning (1)
- latches (1)
- learning (1)
- leftmost derivations (1)
- lesson planning (1)
- lesson preparation (1)
- linear programming (1)
- logic programming (1)
- logic programming methodology and applications (1)
- logic-based modeling (1)
- loop formulas (1)
- loose programming (1)
- loss propagation (1)
- lunar exploration (1)
- machine learning (1)
- machine learning algorithms (1)
- manipulation planning (1)
- measure development (1)
- media (1)
- mediated learning experience (1)
- metabolic network (1)
- metabolism (1)
- metabolomics (1)
- metadata (1)
- metastasis (1)
- mobile learning (1)
- mobile technologies and apps (1)
- mobiles Lernen (1)
- natural disasters (1)
- natural language generation (1)
- neighborhood (1)
- networks (1)
- neural networks (1)
- neutral endopeptidase (1)
- nonphotorealistic rendering (NPR) (1)
- o-ambiguity (1)
- on-farm evaluation (1)
- open learning (1)
- operating system (1)
- organisational evolution (1)
- pMOS radiation dosimeter (1)
- paper prototyping (1)
- parallel processing (1)
- parallel rewriting (1)
- parameter (1)
- parity aggregate operator (1)
- parsing (1)
- pdf forms (1)
- pedagogy (1)
- perception (1)
- perception differences (1)
- personal (1)
- personal response systems (1)
- philosophical foundation of informatics pedagogy (1)
- physical computing (1)
- physical computing tools (1)
- planning (1)
- platform (1)
- plug-ins (1)
- policy evaluation (1)
- policy making (1)
- portfolio-based solving (1)
- pre-primary level (1)
- predictive models (1)
- premise acceptability (1)
- preprocessing (1)
- primary education (1)
- primary level (1)
- problem-solving (1)
- process model alignment (1)
- process modeling (1)
- professional development (1)
- program encodings (1)
- programmed grammars (1)
- programming (1)
- programming in context (1)
- projection (1)
- proof complexity (1)
- pruritus (1)
- pulse stretching inverters (1)
- quality of life (1)
- quantum (1)
- random forest (1)
- real arguments (1)
- real-time (1)
- real-time mapping (1)
- reference (1)
- referential effectiveness (1)
- regression (1)
- regular language (1)
- regulation (1)
- reliability (1)
- reliability analysis (1)
- resources (1)
- restricted parallelism (1)
- satisfiability (1)
- secondary computer science education (1)
- secondary education (1)
- selective fault tolerance (1)
- self-adaptive multiprocessing system (1)
- self-checking (1)
- self-efficacy (1)
- semantic web (1)
- sensitivity (1)
- simplicity (1)
- single event transients (1)
- single event upset (1)
- single event upsets (1)
- single-event transient (1)
- single-trial-analysis (1)
- site-specific weed management (1)
- sleep quality (1)
- smart farming (1)
- social media (1)
- solar particle event (1)
- space missions (1)
- stable model semantics (1)
- state complexity (1)
- static analysis (1)
- static prediction games (1)
- strong equivalence (1)
- student activation (1)
- student experience (1)
- student perceptions (1)
- students’ conceptions (1)
- students’ knowledge (1)
- sufficiency (1)
- suicidal ideations (1)
- supply chains (1)
- support system (1)
- support vector machines (1)
- tableau calculi (1)
- teacher competencies (1)
- teacher training (1)
- teaching (1)
- teaching informatics in general education (1)
- technical notes and rapid communications (1)
- technische Rahmenbedingungen (1)
- tele-teaching (1)
- test response compaction (1)
- theorem (1)
- theory (1)
- theory of computation (1)
- timing (1)
- tracing (1)
- transient Faults (1)
- transient analysis (1)
- triangulated irregular networks (1)
- triple modular redundancy (1)
- unfounded sets (1)
- unification (1)
- user experience (1)
- user-centred (1)
- verification (1)
- video annotation (1)
- virtual mobility (1)
- virtual reality (1)
- wheat crops (1)
- work productivity (1)
- xAPI (1)
- yellow rust (1)
- ‘unplugged’ computing (1)
Institute
- Institut für Informatik und Computational Science (682)
- Institut für Physik und Astronomie (2)
- eLiS - E-Learning in Studienbereichen (2)
- Department Erziehungswissenschaft (1)
- Department Linguistik (1)
- Extern (1)
- Fachgruppe Betriebswirtschaftslehre (1)
- Historisches Institut (1)
- Universitätsbibliothek (1)
- Zentrum für Qualitätsentwicklung in Lehre und Studium (ZfQ) (1)
Im Bachelor-Studiengang (B. Sc.) IT Security an der Fachhochschule St. Pölten wurde im Wintersemester 2011/12 versuchsweise die Lehrorganisation im ersten Fachsemester verändert: Die Module bzw. Teilmodule wurden nicht mehr alle parallel zueinander unterrichtet, sondern jedes Modul wurde exklusiv über einige Wochen abgehalten. Im Beitrag werden die Auswirkungen und bisherigen Erfahrungen mit dieser Reorganisation der Lehre geschildert: So haben sich die Noten im Mittel um etwa eine Note verbessert, die Zahl derjenigen Studierenden, die durch Prüfungen durchfallen, ist drastisch gesunken. Die Zufriedenheit der Studierenden und Lehrenden ist so groß, dass diese Form der Lehrorganisation im gesamten Bachelor- und auch im Masterstudiengang übernommen wird.
Eine übliche Erzählung verknüpft lange Studienzeiten und hohe Abbrecherquoten im Informatikstudium zum einen mit der sehr gut bezahlten Nebentätigkeit von Studierenden in der Informatikbranche, die deutlich studienzeitverlängernd sei; zum anderen werde wegen des hohen Bedarfs an Informatikern ein formeller Studienabschluss von den Studierenden häufig als entbehrlich betrachtet und eine Karriere in der Informatikbranche ohne abgeschlossenes Studium begonnen. In dieser Studie, durchgeführt an der Universität Potsdam, untersuchen wir, wie viele Informatikstudierende neben dem Studium innerhalb und außerhalb der Informatikbranche arbeiten, welche Erwartungen sie neben der Bezahlung damit verbinden und wie sich die Tätigkeit auf ihr Studium und ihre spätere berufliche Perspektive auswirkt. Aus aktuellem Anlass interessieren uns auch die Auswirkungen der Covid-19-Pandemie auf die Arbeitstätigkeiten der Informatikstudierenden.
Using the timing flexibility of modern automatic test equipment (ATE) test response data can be compacted without the need for additional X-masking logic. In this article the test response is compacted by several multiple input shift registers without feedback (NF-MISR). The shift registers are running on a k-times higher clock frequency than the test clock. For each test clock cycle only one out of the k outputs of each shift register is evaluated by the ATE. The impact of consecutive X values within the scan chains is reduced by a periodic permutation of the NF-MISR inputs. As a result, no additional external control signals or test set dependent control logic is required. The benefits of the proposed method are shown by the example of an implementation on a Verigy ATE. Experiments on three industrial circuits demonstrate the effectiveness of the proposed approach in comparison to a commercial DFT solution.
Analyses of metagenomes in life sciences present new opportunities as well as challenges to the scientific community and call for advanced computational methods and workflows. The large amount of data collected from samples via next-generation sequencing (NGS) technologies render manual approaches to sequence comparison and annotation unsuitable. Rather, fast and efficient computational pipelines are needed to provide comprehensive statistics and summaries and enable the researcher to choose appropriate tools for more specific analyses. The workflow presented here builds upon previous pipelines designed for automated clustering and annotation of raw sequence reads obtained from next-generation sequencing technologies such as 454 and Illumina. Employing specialized algorithms, the sequence reads are processed at three different levels. First, raw reads are clustered at high similarity cutoff to yield clusters which can be exported as multifasta files for further analyses. Independently, open reading frames (ORFs) are predicted from raw reads and clustered at two strictness levels to yield sets of non-redundant sequences and ORF families. Furthermore, single ORFs are annotated by performing searches against the Pfam database
With the success of wireless technologies in consumer electronics, standard wireless technologies are envisioned for the deployment in industrial environments as well. Industrial applications involving mobile subsystems or just the desire to save cabling make wireless technologies attractive. Nevertheless, these applications often have stringent requirements on reliability and timing. In wired environments, timing and reliability are well catered for by fieldbus systems (which are a mature technology designed to enable communication between digital controllers and the sensors and actuators interfacing to a physical process). When wireless links are included, reliability and timing requirements are significantly more difficult to meet, due to the adverse properties of the radio channels. In this paper we thus discuss some key issues coming up in wireless fieldbus and wireless industrial communication systems:1)fundamental problems like achieving timely and reliable transmission despite channel errors; 2) the usage of existing wireless technologies for this specific field of applications; and 3) the creation of hybrid systems in which wireless stations are included into existing wired systems
A survey has been carried out in the Computer Science (CS) department at the University of Baghdad to investigate the attitudes of CS students in a female dominant environment, showing the differences between male and female students in different academic years. We also compare the attitudes of the freshman students of two different cultures (University of Baghdad, Iraq, and the University of Potsdam).
A lot has been published about the competencies needed by
students in the 21st century (Ravenscroft et al., 2012). However, equally
important are the competencies needed by educators in the new era
of digital education. We review the key competencies for educators in
light of the new methods of teaching and learning proposed by Massive
Open Online Courses (MOOCs) and their on-campus counterparts,
Small Private Online Courses (SPOCs).
Der Unterricht großer Studierendengruppen im wissenschaftlichen Schreiben birgt vielfältige organisatorische Herausforderungen und eine zeitintensive Betreuung durch die Dozenten. Diese Arbeit stellt ein Lehrkonzept mit Peer-Reviews vor, in dem das Feedback der Peers durch eine automatisierte Analyse ergänzt wird. Die Software Confopy liefert metrik- und strukturbasierte Hinweise für die Verbesserung des wissenschaftlichen Schreibstils. Der Nutzen von Confopy wird an 47 studentischen Arbeiten in Draft- und Final-Version illustriert.
Geometric generalization is a fundamental concept in the digital mapping process. An increasing amount of spatial data is provided on the web as well as a range of tools to process it. This jABC workflow is used for the automatic testing of web-based generalization services like mapshaper.org by executing its functionality, overlaying both datasets before and after the transformation and displaying them visually in a .tif file. Mostly Web Services and command line tools are used to build an environment where ESRI shapefiles can be uploaded, processed through a chosen generalization service and finally visualized in Irfanview.
Tropical cyclones range among the costliest disasters on Earth. Their economic repercussions along the supply and trade network also affect remote economies that are not directly affected. We here simulate possible global repercussions on consumption for the example case of Hurricane Sandy in the US (2012) using the shock-propagation model Acclimate. The modeled shock yields a global three-phase ripple: an initial production demand reduction and associated consumption price decrease, followed by a supply shortage with increasing prices, and finally a recovery phase. Regions with strong trade relations to the US experience strong magnitudes of the ripple. A dominating demand reduction or supply shortage leads to overall consumption gains or losses of a region, respectively. While finding these repercussions in historic data is challenging due to strong volatility of economic interactions, numerical models like ours can help to identify them by approaching the problem from an exploratory angle, isolating the effect of interest. For this, our model simulates the economic interactions of over 7000 regional economic sectors, interlinked through about 1.8 million trade relations. Under global warming, the wave-like structures of the economic response to major hurricanes like the one simulated here are likely to intensify and potentially overlap with other weather extremes.
Die Wahl des richtigen Studienfaches und die daran anschließende
Studieneingangsphase sind oft entscheidend für den erfolgreichen Verlauf eines Studiums. Eine große Herausforderung besteht dabei darin, bereits in den ersten Wochen des Studiums bestehende Defizite in vermeintlich einfachen Schlüsselkompetenzen zu erkennen und diese so bald wie möglich zu beheben. Eine zweite, nicht minder wichtige Herausforderung ist es, möglichst frühzeitig für jeden einzelnen Studierenden zu erkennen, ob er bzw. sie das individuell richtige Studienfach gewählt hat, das den jeweiligen persönlichen Neigungen, Interessen und Fähigkeiten entspricht und zur Verwirklichung der eigenen Lebensziele beiträgt. Denn nur dann sind Studierende ausreichend stark und dauerhaft intrinsisch motiviert, um ein anspruchsvolles, komplexes Studium erfolgreich durchzuziehen. In diesem Beitrag fokussieren wir eine Maßnahme, die die Studierenden an einen Prozess zur systematischen Reflexion des eigenen Lernprozesses und der eigenen Ziele heranführt und beides in Relation setzt.
Was ist eine kreative Leistung von Studierenden? Dies ist die Ausgangsfrage, wenn Lehre kreativitätsförderlicher als bislang gestaltet werden soll. In diesem Beitrag wird ein Modell zur Förderung von Kreativität in der Hochschullehre vorgestellt und mit einem Beispiel verdeutlicht. Es wird die veränderte Konzeption der Vorlesung Informatik & Gesellschaft illustriert: Studierende hatten die Aufgabe, eine „e-Infrastruktur für die Universität NeuDoBoDu“ zu entwickeln. Hierzu werden die Ergebnisse der Evaluation und Erfahrungen erläutert.
Viele Hochschulen nutzen SAP ERP in der Lehre, um den Studierenden einen Einblick in die Funktionsweise und den Aufbau von integrierter Standardsoftware zu ermöglichen. Im Rahmen solcher Schulungen bilden die Studierenden eine Meinung und Bewertung der Software. In diesem Artikel wird untersucht, wie sich klassische Modelle der Nutzungswahrnehmung auf die spezielle Situation von SAP ERP in der Lehre übertragen lassen und welchen Einfluss bestimmte Faktoren haben. Dazu wurden vier Vorher-Nachher-Studien durchgeführt. Die Ergebnisse zeigen, dass die Funktionalität im Laufe der Schulung positiver und die Benutzungsfreundlichkeit als negativer bewertet wird.
Die Studienanfänger der Informatik haben in Deutschland sehr unterschiedliche Grundkenntnisse in der Programmierung. Dies führt immer wieder zu Schwierigkeiten in der Ausrichtung der Einführungsveranstaltungen. An der TU München wird seit dem Wintersemester 2008/2009 nun eine neue Art von Vorkursen angeboten. In nur 2,5 Tagen erstellen die Teilnehmer ein kleines objektorientiertes Programm. Dabei arbeiten sie weitestgehend alleine, unterstützt von einem studentischen Tutor. In dieser Arbeit sollen nun das Konzept der sogenannten „Vorprojekte“ sowie erste Forschungsansätze vorgestellt werden
A workflow for visualizing server connections using the Google Maps API was built in the jABC. It makes use of three basic services: An XML-based IP address geolocation web service, a command line tool and the Static Maps API. The result of the workflow is an URL leading to an image file of a map, showing server connections between a client and a target host.
Recent philosophical analyses of the epistemic dimension of images in the sciences show a certain trend in acknowledging potential roles of these images beyond their merely decorative or pedagogical functions. We argue, however, that this new debate has yet paid little attention to a special type of pictures, we call ‘visual metaphor’, and its versatile heuristic potential in organizing data, supporting communication, and guiding research, modeling, and theory formation. Based on a case study of Conrad Hal Waddington’s epigenetic landscape images in biology, we develop a descriptive framework applicable to heuristic roles of various visual metaphors in the sciences.
This paper continues the line of research aimed at investigating the relationship between logic programs and first-order theories. We extend the definition of program completion to programs with input and output in a subset of the input language of the ASP grounder gringo, study the relationship between stable models and completion in this context, and describe preliminary experiments with the use of two software tools, anthem and vampire, for verifying the correctness of programs with input and output. Proofs of theorems are based on a lemma that relates the semantics of programs studied in this paper to stable models of first-order formulas.
Peer Assessment ist eine Methode, bei der die Teilnehmer eine gestellte Aufgabe nicht nur bearbeiten und einreichen, sondern – in einer zweiten Phase – diese auch gegenseitig überprüfen, kommentieren und bewerten. Durch diese Methode wird, auch in sehr großen Veranstaltungen, das Üben mit individuellen Bewertungen und individuellem Feedback möglich.
Im Wintersemester 2013/14 wurde dieser Ansatz in der Erstsemesterveranstaltung Programmieren an der Technischen Hochschule Nürnberg mit 340 Studierenden als semesterbegleitendes Online-Pflichtpraktikum erprobt. Bei gleichen Leistungsanforderungen wurde bei Studierenden, die erfolgreich am Praktikum teilnahmen, eine Reduzierung der Durchfallquote um durchschnittlich 60 % und eine Verbesserung der Durchschnittsnote um 0,6 – 0,9 Notenstufen erzielt. Zudem lernten die teilnehmenden Studierenden kontinuierlicher, bereiteten Lerninhalte besser nach und gelangten zu einer überwiegend positiven Einschätzung des Praktikums und der Methode. Im E-Learning System Moodle kann Peer Assessment, mit moderatem Umsetzungs- und Betreuungsaufwand, mit der Workshop-Aktivität realisiert werden. Im Beitrag wird auf die Schlüsselelemente des erfolgreichen Einsatzes von Peer Assessment eingegangen.
In this paper we introduce and study some new cooperation protocols for cooperating distributed (CD) grammar systems. These derivation modes depend on the number of different nonterminals present in the sentential form obtained when a component finished a derivation phase. This measure describes the competence of the grammar on the string (the competence is high if the number of the different nonterminals is small). It is also a measure of the efficiency of the grammar on the given string (a component is more efficient than another one if it is able to decrease the number of nonterminals in the string to a greater extent). We prove that if the underlying derivation mode is the t-mode derivation, then some variants of these systems determine the class of random context ET0L languages. If these CD grammar systems use the k step limited derivations as underlying derivation mode, then they are able to generate any recursively enumerable language.
Background: The biological interpretation of large-scale gene expression data is one of the paramount challenges in current bioinformatics. In particular, placing the results in the context of other available functional genomics data, such as existing bio-ontologies, has already provided substantial improvement for detecting and categorizing genes of interest. One common approach is to look for functional annotations that are significantly enriched within a group or cluster of genes, as compared to a reference group. Results: In this work, we suggest the information-theoretic concept of mutual information to investigate the relationship between groups of genes, as given by data-driven clustering, and their respective functional categories. Drawing upon related approaches (Gibbons and Roth, Genome Research 12: 1574-1581, 2002), we seek to quantify to what extent individual attributes are sufficient to characterize a given group or cluster of genes. Conclusion: We show that the mutual information provides a systematic framework to assess the relationship between groups or clusters of genes and their functional annotations in a quantitative way. Within this framework, the mutual information allows us to address and incorporate several important issues, such as the interdependence of functional annotations and combinatorial combinations of attributes. It thus supplements and extends the conventional search for overrepresented attributes within a group or cluster of genes. In particular taking combinations of attributes into account, the mutual information opens the way to uncover specific functional descriptions of a group of genes or clustering result. All datasets and functional annotations used in this study are publicly available. All scripts used in the analysis are provided as additional files.
Scientific writing is an important skill for computer science and computer engineering professionals. In this paper we present a writing concept across the curriculum program directed towards scientific writing. The program is built around a hierarchy of learning outcomes. The hierarchy is constructed through analyzing the learning outcomes in relation to competencies that are needed to fulfill them.
Current curricular trends require teachers in Baden-
Wuerttemberg (Germany) to integrate Computer Science (CS) into
traditional subjects, such as Physical Science. However, concrete guidelines
are missing. To fill this gap, we outline an approach where a
microcontroller is used to perform and evaluate measurements in the
Physical Science classroom.
Using the open-source Arduino platform, we expect students to acquire
and develop both CS and Physical Science competencies by using a
self-programmed microcontroller. In addition to this combined development
of competencies in Physical Science and CS, the subject matter
will be embedded in suitable contexts and learning environments,
such as weather and climate.
User Experience (UX) describes the holistic experience of a user before, during, and after interaction with a platform, product, or service. UX adds value and attraction to their sole functionality and is therefore highly relevant for firms. The increased interest in UX has produced a vast amount of scholarly research since 1983. The research field is, therefore, complex and scattered. Conducting a bibliometric analysis, we aim at structuring the field quantitatively and rather abstractly. We employed citation analyses, co-citation analyses, and content analyses to evaluate productivity and impact of extant research. We suggest that future research should focus more on business and management related topics.
Untitled
(2004)
In diesem Papier wird das Konzept eines Lernzentrums für die Informatik (LZI) an der Universität Paderborn vorgestellt. Ausgehend von den fachspezifischen Schwierigkeiten der Informatik Studierenden werden die Angebote des LZIs erläutert, die sich über die vier Bereiche Individuelle Beratung und Betreuung, „Offener Lernraum“, Workshops und Lehrveranstaltungen sowie Forschung erstrecken. Eine erste Evaluation mittels Feedbackbögen zeigt, dass das Angebot bei den Studierenden positiv aufgenommen wird. Zukünftig soll das Angebot des LZIs weiter ausgebaut und verbessert werden. Ausgangsbasis dazu sind weitere Studien.
We investigate the operation problem for linear and deterministic context-free languages: Fix an operation on formal languages. Given linear (deterministic, respectively) context-free languages, is the application of this operation to the given languages still a linear (deterministic, respectively) context-free language? Besides the classical operations, for which the linear and deterministic context-free languages are not closed, we also consider the recently introduced root and power operation. We show non-semidecidability, to be more precise, we show completeness for the second level of the arithmetic hierarchy for all of the aforementioned operations, except for the power operation, if the underlying alphabet contains at least two letters. The result for the power opera, tion solves an open problem stated in Theoret. Comput. Sci. 314 (2004) 445-449
Computational thinking is a fundamental skill set that is learned
by studying Informatics and ICT. We argue that its core ideas can
be introduced in an inspiring and integrated way to both teachers and
students using fun and contextually rich cs4fn ‘Computer Science for
Fun’ stories combined with ‘unplugged’ activities including games and
magic tricks. We also argue that understanding people is an important
part of computational thinking. Computational thinking can be fun for
everyone when taught in kinaesthetic ways away from technology.
Unmixing hyperspectral data
(2000)
Parallel communicating finite automata (PCFAs) are systems of several finite state automata which process a common input string in a parallel way and are able to communicate by sending their states upon request. We consider deterministic and nondeterministic variants and distinguish four working modes. It is known that these systems in the most general mode are as powerful as one-way multi-head finite automata. It is additionally known that the number of heads corresponds to the number of automata in PCFAs in a constructive way. Thus, undecidability results as well as results on the hierarchies induced by the number of heads carry over from multi-head finite automata to PCFAs in the most general mode. Here, we complement these undecidability and hierarchy results also for the remaining working modes. In particular, we show that classical decidability questions are not semi-decidable for any type of PCFAs under consideration. Moreover, it is proven that the number of automata in the system induces infinite hierarchies for deterministic and nondeterministic PCFAs in three working modes.
Berufsbegleitende Studiengänge stehen vor besonderen Schwierigkeiten, für die der Einsatz von Blended Learning-Szenarien sinnvoll sein kann. Welche speziellen Herausforderungen sich dabei ergeben und welche Lösungsansätze dagegen steuern, betrachtet der folgende Artikel anhand eines Praxisberichts aus dem Studiengang M. P. A. Wissenschaftsmanagement an der Universität Speyer.
The UDKM1DSIM toolbox is a collection of MATLAB (MathWorks Inc.) classes and routines to simulate the structural dynamics and the according X-ray diffraction response in one-dimensional crystalline sample structures upon an arbitrary time-dependent external stimulus, e.g. an ultrashort laser pulse. The toolbox provides the capabilities to define arbitrary layered structures on the atomic level including a rich database of corresponding element-specific physical properties. The excitation of ultrafast dynamics is represented by an N-temperature model which is commonly applied for ultrafast optical excitations. Structural dynamics due to thermal stress are calculated by a linear-chain model of masses and springs. The resulting X-ray diffraction response is computed by dynamical X-ray theory. The UDKM1DSIM toolbox is highly modular and allows for introducing user-defined results at any step in the simulation procedure.
Program summary
Program title: udkm1Dsim
Catalogue identifier: AERH_v1_0
Program summary URL: http://cpc.cs.qub.ac.uk/summaries/AERH_v1_0.html
Licensing provisions: BSD
No. of lines in distributed program, including test data, etc.: 130221
No. of bytes in distributed program, including test data, etc.: 2746036
Distribution format: tar.gz
Programming language: Matlab (MathWorks Inc.).
Computer: PC/Workstation.
Operating system: Running Matlab installation required (tested on MS Win XP -7, Ubuntu Linux 11.04-13.04).
Has the code been vectorized or parallelized?: Parallelization for dynamical XRD computations. Number of processors used: 1-12 for Matlab Parallel Computing Toolbox; 1 - infinity for Matlab Distributed Computing Toolbox
External routines:
Optional: Matlab Parallel Computing Toolbox, Matlab Distributed Computing Toolbox Required (included in the package): mtimesx Fast Matrix Multiply for Matlab by James Tursa, xml io tools by Jaroslaw Tuszynski, textprogressbar by Paul Proteus
Nature of problem:
Simulate the lattice dynamics of 1D crystalline sample structures due to an ultrafast excitation including thermal transport and compute the corresponding transient X-ray diffraction pattern.
Solution method:
Restrictions:
The program is restricted to 1D sample structures and is further limited to longitudinal acoustic phonon modes and symmetrical X-ray diffraction geometries.
Unusual features: The program is highly modular and allows the inclusion of user-defined inputs at any time of the simulation procedure.
Running time: The running time is highly dependent on the number of unit cells in the sample structure and other simulation parameters such as time span or angular grid for X-ray diffraction computations. However, the example files are computed in approx. 1-5 min each on a 8 Core Processor with 16 GB RAM available.
Answer Set Programming (ASP) is a paradigm for modeling and solving problems for knowledge representation and reasoning. There are plenty of results dedicated to studying the hardness of (fragments of) ASP. So far, these studies resulted in characterizations in terms of computational complexity as well as in fine-grained insights presented in form of dichotomy-style results, lower bounds when translating to other formalisms like propositional satisfiability (SAT), and even detailed parameterized complexity landscapes. A generic parameter in parameterized complexity originating from graph theory is the socalled treewidth, which in a sense captures structural density of a program. Recently, there was an increase in the number of treewidth-based solvers related to SAT. While there are translations from (normal) ASP to SAT, no reduction that preserves treewidth or at least keeps track of the treewidth increase is known. In this paper we propose a novel reduction from normal ASP to SAT that is aware of the treewidth, and guarantees that a slight increase of treewidth is indeed sufficient. Further, we show a new result establishing that, when considering treewidth, already the fragment of normal ASP is slightly harder than SAT (under reasonable assumptions in computational complexity). This also confirms that our reduction probably cannot be significantly improved and that the slight increase of treewidth is unavoidable. Finally, we present an empirical study of our novel reduction from normal ASP to SAT, where we compare treewidth upper bounds that are obtained via known decomposition heuristics. Overall, our reduction works better with these heuristics than existing translations. (c) 2021 Elsevier B.V. All rights reserved.
Programs are often subjected to significant optimizing and parallelizing transformations based on extensive dependence analysis. Formal validation of such transformations needs modelling paradigms which can capture both control and data dependences in the program vividly. Being value-based with an inherent scope of capturing parallelism, the untimed coloured Petri net (CPN) models, reported in the literature, fit the bill well; accordingly, they are likely to be more convenient as the intermediate representations (IRs) of both the source and the transformed codes for translation validation than strictly sequential variable-based IRs like sequential control flow graphs (CFGs). In this work, an efficient path-based equivalence checking method for CPN models of programs on integers is presented. Extensive experimentation has been carried out on several sequential and parallel examples. Complexity and correctness issues have been treated rigorously for the method.
A well-known result by Stein (1956) shows that in particular situations, biased estimators can yield better parameter estimates than their generally preferred unbiased counterparts. This letter follows the same spirit, as we will stabilize the unbiased generalization error estimates by regularization and finally obtain more robust model selection criteria for learning. We trade a small bias against a larger variance reduction, which has the beneficial effect of being more precise on a single training set. We focus on the subspace information criterion (SIC), which is an unbiased estimator of the expected generalization error measured by the reproducing kernel Hilbert space norm. SIC can be applied to the kernel regression, and it was shown in earlier experiments that a small regularization of SIC has a stabilization effect. However, it remained open how to appropriately determine the degree of regularization in SIC. In this article, we derive an unbiased estimator of the expected squared error, between SIC and the expected generalization error and propose determining the degree of regularization of SIC such that the estimator of the expected squared error is minimized. Computer simulations with artificial and real data sets illustrate that the proposed method works effectively for improving the precision of SIC, especially in the high-noise-level cases. We furthermore compare the proposed method to the original SIC, the cross-validation, and an empirical Bayesian method in ridge parameter selection, with good results
We introduce a type and effect system, for an imperative object calculus, which infers sharing possibly introduced by the evaluation of an expression, represented as an equivalence relation among its free variables. This direct representation of sharing effects at the syntactic level allows us to express in a natural way, and to generalize, widely-used notions in literature, notably uniqueness and borrowing. Moreover, the calculus is pure in the sense that reduction is defined on language terms only, since they directly encode store. The advantage of this non-standard execution model with respect to a behaviorally equivalent standard model using a global auxiliary structure is that reachability relations among references are partly encoded by scoping. (C) 2018 Elsevier B.V. All rights reserved.
The submission and management of computational jobs is a traditional part of utility computing environments. End users and developers of domain-specific software abstractions often have to deal with the heterogeneity of such batch processing systems. This lead to a number of application programming interface and job description standards in the past, which are implemented and established for cluster and Grid systems. With the recent rise of cloud computing as new utility computing paradigm, the standardized access to batch processing facilities operated on cloud resources becomes an important issue. Furthermore, the design of such a standard has to consider a tradeoff between feature completeness and the achievable level of interoperability. The article discusses this general challenge, and presents some existing standards with traditional cluster and Grid computing background that may be applicable to cloud environments. We present OCCI-DRMAA as one approach for standardized access to batch processing facilities hosted in a cloud.
We elaborate upon the theoretical foundations of a metric temporal extension of Answer Set Programming. In analogy to previous extensions of ASP with constructs from Linear Temporal and Dynamic Logic, we accomplish this in the setting of the logic of Here-and-There and its non-monotonic extension, called Equilibrium Logic. More precisely, we develop our logic on the same semantic underpinnings as its predecessors and thus use a simple time domain of bounded time steps. This allows us to compare all variants in a uniform framework and ultimately combine them in a common implementation.
Answer Set Programming (ASP) is a successful rule-based formalism for modeling and solving knowledge-intense combinatorial (optimization) problems. Despite its success in both academic and industry, open challenges like automatic source code optimization, and software engineering remains. This is because a problem encoded into an ASP might not have the desired solving performance compared to an equivalent representation. Motivated by these two challenges, this paper has three main contributions. First, we propose a developing process towards a methodology to implement ASP programs, being faithful to existing methods. Second, we present ASP encodings that serve as the basis from the developing process. Third, we demonstrate the use of ASP to reverse the standard solving process. That is, knowing answer sets in advance, and desired strong equivalent properties, “we” exhaustively reconstruct ASP programs if they exist. This paper was originally motivated by the search of propositional formulas (if they exist) that represent the semantics of a new aggregate operator. Particularly, a parity aggregate. This aggregate comes as an improvement from the already existing parity (xor) constraints from xorro, where lacks expressiveness, even though these constraints fit perfectly for reasoning modes like sampling or model counting. To this end, this extended version covers the fundaments from parity constraints as well as the xorro system. Hence, we delve a little more in the examples and the proposed methodology over parity constraints. Finally, we discuss our results by showing the only representation available, that satisfies different properties from the classical logic xor operator, which is also consistent with the semantics of parity constraints from xorro.
Non-stationarities are ubiquitous in EEG signals. They are especially apparent in the use of EEG-based brain- computer interfaces (BCIs): (a) in the differences between the initial calibration measurement and the online operation of a BCI, or (b) caused by changes in the subject's brain processes during an experiment (e.g. due to fatigue, change of task involvement, etc). In this paper, we quantify for the first time such systematic evidence of statistical differences in data recorded during offline and online sessions. Furthermore, we propose novel techniques of investigating and visualizing data distributions, which are particularly useful for the analysis of (non-) stationarities. Our study shows that the brain signals used for control can change substantially from the offline calibration sessions to online control, and also within a single session. In addition to this general characterization of the signals, we propose several adaptive classification schemes and study their performance on data recorded during online experiments. An encouraging result of our study is that surprisingly simple adaptive methods in combination with an offline feature selection scheme can significantly increase BCI performance
Novel two-dimensional tactile displays enable blind users to not only get access to the textual but also to the graphical content of a graphical user interface. Due to the higher amount of information that can be presented in parallel, orientation and exploration can be more complex. In this paper we present the HyperBraille system, which consists of a pin-matrix device as well as a graphical screen reader providing the user with appropriate presentation and interaction possibilities. To allow for a detailed analysis of bimanual interaction strategies on a pin-matrix device, we conducted two user studies with a total of 12 blind people. The task was to fill in .pdf forms on the pin-matrix device by using different input methods, namely gestures, built-in hardware buttons as well as a conventional PC keyboard. The forms were presented in a semigraphic view type that not only contains Braille but also tactile widgets in a spatial arrangement. While completion time and error rate partly depended on the chosen input method, the usage of special reading strategies seemed to be independent of it. A direct comparison of the system and a conventional assistive technology (screen reader with single-line Braille device) showed that interaction on the pin-matrix device can be very efficient if the user is trained. The two-dimensional output can improve access to .pdf forms with insufficient accessibility as the mapping of input controls and the corresponding labels can be supported by a spatial presentation.
Motivation: Continued development of analytical techniques based on gas chromatography and mass spectrometry now facilitates the generation of larger sets of metabolite concentration data. An important step towards the understanding of metabolite dynamics is the recognition of stable states where metabolite concentrations exhibit a simple behaviour. Such states can be characterized through the identification of significant thresholds in the concentrations. But general techniques for finding discretization thresholds in continuous data prove to be practically insufficient for detecting states due to the weak conditional dependences in concentration data. Results: We introduce a method of recognizing states in the framework of decision tree induction. It is based upon a global analysis of decision forests where stability and quality are evaluated. It leads to the detection of thresholds that are both comprehensible and robust. Applied to metabolite concentration data, this method has led to the discovery of hidden states in the corresponding variables. Some of these reflect known properties of the biological experiments, and others point to putative new states
Think logarithmically!
(2015)
We discuss here a number of algorithmic topics which we
use in our teaching and in learning of mathematics and informatics to
illustrate and document the power of logarithm in designing very efficient
algorithms and computations – logarithmic thinking is one of the
most important key competencies for solving real world practical problems.
We demonstrate also how to introduce logarithm independently
of mathematical formalism using a conceptual model for reducing a
problem size by at least half. It is quite surprising that the idea, which
leads to logarithm, is present in Euclid’s algorithm described almost
2000 years before John Napier invented logarithm.
Die Möglichkeiten sich zu informieren, am Leben der vielen Anderen teilzunehmen ist durch das Internet mit seinen Tweets, Google-Angeboten und sozialen Netzwerken wie Facebook ins Unermessliche gewachsen. Zugleich fühlen sich viele Nutzer überfordert und meinen, im Meer der Informationen zu ertrinken. So bekennt Frank Schirrmacher in seinem Buch Payback, dass er den geistigen Anforderungen unserer Zeit nicht mehr gewachsen ist. Sein Kopf komme nicht mehr mit. Er sei unkonzentriert, vergesslich und ständig abgelenkt. Das, was vielen zum Problem geworden ist, sehen viele Studierende eher pragmatisch. Der Wissenserwerb in Zeiten von Internet und E-Learning läuft an Hochschulen häufig nach der Helene-Hegemann-Methode ab: Zunächst machen sich die Studierenden, z.B. im Rahmen einer Studien- oder Hausarbeit, bei Wikipedia „schlau“, ein Einstieg ist geschafft. Anschließend wird dieses Wissen mit Google angereichert. Damit ist Überblickswissen vorhanden. Mit geschickter copy-and-paste-Komposition lässt sich daraus schon ein „Werk“ erstellen. Der ein oder andere Studierende gibt sich mit diesem Wissenserwerb zufrieden und bricht seinen Lernprozess hier bereits ab. Nun ist zwar am Ende jeder Studierende für seinen Wissenserwerb selbst verantwortlich. Die erkennbar unbefriedigende Situation sollte die Hochschulen aber herausfordern, das Internet in Vorlesungen und Seminaren auszuprobieren und sinnvolle Anwendungen zu entwickeln. Beispiele gibt es durchaus. Unter der Metapher E-Learning hat sich ein umfangreicher Forschungsschwerpunkt an den Universitäten entwickelt. Einige Beispiele von vielen: So hat der Osnabrücker Informatik-Professor Oliver Vornberger seine Vorlesungen als Video ins Netz gestellt. Per RSS ist es möglich, Sequenzen aufs iPod zu laden. Die übliche Dozentenangst, dann würden sie ja vor leeren Bänken sitzen, scheint unbegründet. Sie werden von den Studierenden vor allem zur Prüfungsvorbereitung genutzt. Wie ist das Internet, das für die junge Generation zu einem alles andere verdrängenden Universalmedium geworden ist, didaktisch in die Hochschullehre einzubinden? Wie also ist konkret mit diesen Herausforderungen umzugehen? Dies soll uns im Folgenden beschäftigen.
The Technology Proficiency Self-Assessment (TPSA) questionnaire
has been used for 15 years in the USA and other nations as a
self-efficacy measure for proficiencies fundamental to effective technology
integration in the classroom learning environment. Internal consistency
reliabilities for each of the five-item scales have typically ranged
from .73 to .88 for preservice or inservice technology-using teachers.
Due to changing technologies used in education, researchers sought to
renovate partially obsolete items and extend self-efficacy assessment to
new areas, such as social media and mobile learning. Analysis of 2014
data gathered on a new, 34 item version of the TPSA indicates that the
four established areas of email, World Wide Web (WWW), integrated
applications, and teaching with technology continue to form consistent
scales with reliabilities ranging from .81 to .93, while the 14 new items
gathered to represent emerging technologies and media separate into
two scales, each with internal consistency reliabilities greater than .9.
The renovated TPSA is deemed to be worthy of continued use in the
teaching with technology context.
The Student Learning Ecology
(2015)
Educational research on social media has showed that
students use it for socialisation, personal communication, and informal
learning. Recent studies have argued that students to some degree use
social media to carry out formal schoolwork. This article gives an
explorative account on how a small sample of Norwegian high school
students use social media to self-organise formal schoolwork. This
user pattern can be called a “student learning ecology”, which is a
user perspective on how participating students gain access to learning
resources.
Answer Set Programming (ASP) is a well-known paradigm of declarative programming with roots in logic programming and non-monotonic reasoning. Similar to other closely related problemsolving technologies, such as SAT/SMT, QBF, Planning and Scheduling, advancements in ASP solving are assessed in competition events. In this paper, we report about the design and results of the Sixth ASP Competition, which was jointly organized by the University of Calabria (Italy), Aalto University (Finland), and the University of Genoa (Italy), in affiliation with the 13th International Conference on Logic Programming and Non-Monotonic Reasoning. This edition maintained some of the design decisions introduced in 2014, e.g., the conception of sub-tracks, the scoring scheme,and the adherence to a fixed modeling language in order to push the adoption of the ASP-Core-2 standard. On the other hand, it featured also some novelties, like a benchmark selection stage classifying instances according to their empirical hardness, and a “Marathon” track where the topperforming systems are given more time for solving hard benchmarks.
Answer Set Programming (ASP) is a prominent knowledge representation language with roots in logic programming and non-monotonic reasoning. Biennial ASP competitions are organized in order to furnish challenging benchmark collections and assess the advancement of the state of the art in ASP solving. In this paper, we report on the design and results of the Seventh ASP Competition, jointly organized by the University of Calabria (Italy), the University of Genova (Italy), and the University of Potsdam (Germany), in affiliation with the 14th International Conference on Logic Programming and Non-Monotonic Reasoning (LPNMR 2017).
The Potsdam answer set solving collection, or Potassco for short, bundles various tools implementing and/or applying answer set programming. The article at hand succeeds an earlier description of the Potassco project published in Gebser et al. (AI Commun 24(2):107-124, 2011). Hence, we concentrate in what follows on the major features of the most recent, fifth generation of the ASP system clingo and highlight some recent resulting application systems.
This article shows a discussion about the key competencies
in informatics and ICT viewed from a philosophical foundation presented
by Martha Nussbaum, which is known as ‘ten central capabilities’.
Firstly, the outline of ‘The Capability Approach’, which has been presented
by Amartya Sen and Nussbaum as a theoretical framework of
assessing the state of social welfare, will be explained. Secondly, the
body of Nussbaum’s ten central capabilities and the reason for being
applied as the basis of discussion will be shown. Thirdly, the relationship
between the concept of ‘capability’ and ‘competency’ is to be
discussed. After that, the author’s assumption of the key competencies
in informatics and ICT led from the examination of Nussbaum’s ten
capabilities will be presented.
The family of default logics
(1998)
The Course's SIB Libraries
(2014)
This chapter gives a detailed description of the service framework underlying all the example projects that form the foundation of this book. It describes the different SIB libraries that we made available for the course “Process modeling in the natural sciences” to provide the functionality that was required for the envisaged applications. The students used these SIB libraries to realize their projects.
We study the complexity of two-person constraint satisfaction games. An instance of such a game is given by a collection of constraints on overlapping sets of variables, and the two players alternately make moves assigning values from a finite domain to the variables, in a specified order. The first player tries to satisfy all constraints, while the other tries to break at least one constraint: the goal is to decide whether the first player has a winning strategy. We show that such games can be conveniently represented by a logical form of quantified constraint satisfaction, where an instance is given by a first-order sentence in which quantifiers alternate and the quantifier-free part is a conjunction of (positive) atomic formulas; the goal is to decide whether the sentence is true. While the problem of deciding such a game is PSPACE-complete in general, by restricting the set of allowed constraint predicates, one can obtain infinite classes of constraint satisfaction games of lower complexity. We use the quantified constraint satisfaction framework to study how the complexity of deciding such a game depends on the parameter set of allowed predicates. With every predicate. one can associate certain predicate-preserving operations, called polymorphisms. We show that the complexity of our games is determined by the surjective polymorphisms of the constraint predicates. We illustrate how this result can be used by identifying the complexity of a wide variety of constraint satisfaction games.
Background:
Pruritus often accompanies chronic skin diseases, exerting considerable burden on many areas of patient functioning; this burden and the features of pruritus remain insufficiently characterized.
Objective:
To investigate characteristics, including localization patterns, and burden of pruritus in patients with chronic dermatoses.
Methods:
We recruited 800 patients with active chronic skin diseases. We assessed pruritus intensity, localization, and further characteristics. We used validated questionnaires to assess quality of life, work productivity and activity impairment, anxiety, depression, and sleep quality.
Results:
Nine out of every 10 patients had experienced pruritus throughout their disease and 73% in the last 7 days. Pruritus often affected the entire body and was not restricted to skin lesions. Patients with moderate to severe pruritus reported significantly more impairment to their sleep quality and work productivity, and they were more depressed and anxious than control individuals and patients with mild or no pruritus. Suicidal ideations were highly prevalent in patients with chronic pruritus (18.5%) and atopic dermatitis (11.8%).
Conclusions:
Pruritus prevalence and intensity are very high across all dermatoses studied; intensity is linked to impairment in many areas of daily functioning. Effective treatment strategies are urgently required to treat pruritus and the underlying skin disease. ( J Am Acad Dermatol 2021;84:691-700.)
The Berlin Brain-Computer Interface (BBCI) project develops a noninvasive BCI system whose key features are 1) the use of well-established motor competences as control paradigms, 2) high-dimensional features from 128-channel electroencephalogram (EEG), and 3) advanced machine learning techniques. As reported earlier, our experiments demonstrate that very high information transfer rates can be achieved using the readiness potential (RP) when predicting the laterality of upcoming left-versus right-hand movements in healthy subjects. A more recent study showed that the RP similarily accompanies phantom movements in arm amputees, but the signal strength decreases with longer loss of the limb. In a complementary approach, oscillatory features are used to discriminate imagined movements (left hand versus right hand versus foot). In a recent feedback study with six healthy subjects with no or very little experience with BCI control, three subjects achieved an information transfer rate above 35 bits per minute (bpm), and further two subjects above 24 and 15 bpm, while one subject could not achieve any BCI control. These results are encouraging for an EEG-based BCI system in untrained subjects that is independent of peripheral nervous system activity and does not rely on evoked potentials even when compared to results with very well-trained subjects operating other BCI systems
A brain-computer interface (BCI) is a system that allows its users to control external devices with brain activity. Although the proof-of-concept was given decades ago, the reliable translation of user intent into device control commands is still a major challenge. Success requires the effective interaction of two adaptive controllers: the user's brain, which produces brain activity that encodes intent, and the BCI system, which translates that activity into device control commands. In order to facilitate this interaction, many laboratories are exploring a variety of signal analysis techniques to improve the adaptation of the BCI system to the user. In the literature, many machine learning and pattern classification algorithms have been reported to give impressive results when applied to BCI data in offline analyses. However, it is more difficult to evaluate their relative value for actual online use. BCI data competitions have been organized to provide objective formal evaluations of alternative methods. Prompted by the great interest in the first two BCI Competitions, we organized the third BCI Competition to address several of the most difficult and important analysis problems in BCI research. The paper describes the data sets that were provided to the competitors and gives an overview of the results.
Interest in developing a new method of man-to-machine communication-a brain-computer interface (BCI)-has grown steadily over the past few decades. BCIs create a new communication channel between the brain and an output device by bypassing conventional motor output pathways of nerves and muscles. These systems use signals recorded from the scalp, the surface of the cortex, or from inside the brain to enable users to control a variety of applications including simple word-processing software and orthotics. BCI technology could therefore provide a new communication and control option for individuals who cannot otherwise express their wishes to the outside world. Signal processing and classification methods are essential tools in the development of improved BCI technology. We organized the BCI Competition 2003 to evaluate the current state of the art of these tools. Four laboratories well versed in EEG-based BCI research provided six data sets in a documented format. We made these data sets (i.e., labeled training sets and unlabeled test sets) and their descriptions available on the Internet. The goal in the competition was to maximize the performance measure for the test labels. Researchers worldwide tested their algorithms and competed for the best classification results. This paper describes the six data sets and the results and function of the most successful algorithms
Testability evaluation of sequential designs incorporating the multi-mode scannable memory element
(1999)
Research publications and data nowadays should be publicly available on the internet and, theoretically, usable for everyone to develop further research, products, or services. The long-term accessibility of research data is, therefore, fundamental in the economy of the research production process. However, the availability of data is not sufficient by itself, but also their quality must be verifiable. Measures to ensure reuse and reproducibility need to include the entire research life cycle, from the experimental design to the generation of data, quality control, statistical analysis, interpretation, and validation of the results. Hence, high-quality records, particularly for providing a string of documents for the verifiable origin of data, are essential elements that can act as a certificate for potential users (customers). These records also improve the traceability and transparency of data and processes, therefore, improving the reliability of results. Standards for data acquisition, analysis, and documentation have been fostered in the last decade driven by grassroot initiatives of researchers and organizations such as the Research Data Alliance (RDA). Nevertheless, what is still largely missing in the life science academic research are agreed procedures for complex routine research workflows. Here, well-crafted documentation like standard operating procedures (SOPs) offer clear direction and instructions specifically designed to avoid deviations as an absolute necessity for reproducibility. Therefore, this paper provides a standardized workflow that explains step by step how to write an SOP to be used as a starting point for appropriate research documentation.