Institut für Informatik und Computational Science
Refine
Has Fulltext
- no (866) (remove)
Year of publication
Document Type
- Article (563)
- Monograph/Edited Volume (133)
- Doctoral Thesis (121)
- Other (28)
- Part of a Book (12)
- Conference Proceeding (5)
- Preprint (4)
Is part of the Bibliography
- yes (866)
Keywords
- Answer set programming (10)
- answer set programming (8)
- Answer Set Programming (7)
- Machine learning (3)
- formal languages (3)
- monitoring (3)
- security (3)
- Analytical models (2)
- Automata systems (2)
- E-learning (2)
- Equilibrium logic (2)
- Event mapping (2)
- Fault tolerance (2)
- Internet (2)
- Internet of Things (2)
- Lindenmayer systems (2)
- MQTT (2)
- Machine Learning (2)
- Modeling (2)
- Non-monotonic reasoning (2)
- Optimization (2)
- Parameterized complexity (2)
- Preference Handling (2)
- Process mining (2)
- ResNet (2)
- TPACK (2)
- Theory (2)
- bioinformatics (2)
- cooperating systems (2)
- e-learning (2)
- knowledge representation and nonmonotonic reasoning (2)
- online learning (2)
- radhard design (2)
- reliability (2)
- verification (2)
- (FPGA) (1)
- (SET) count rate (1)
- 2-tag system (1)
- 3D modeling (1)
- 3D visualization (1)
- AODV (1)
- ASIC (1)
- Absorbed dose (1)
- Abstraction (1)
- Access control (1)
- Active evaluation (1)
- Ad hoc routing (1)
- Adaptivity (1)
- Advanced Video Codec (AVC) (1)
- Aggregates (1)
- Algorithm configuration (1)
- Algorithm portfolios (1)
- Algorithms (1)
- Android Security (1)
- Android hybrid apps (1)
- Animal building (1)
- Answer Set Solving modulo Theories (1)
- Anti-cancer drugs (1)
- App (1)
- Apps (1)
- Argumentation structure (1)
- Artificial Neuronal Network (1)
- Augmentation (1)
- Augmented and virtual reality (1)
- Automated parallelization (1)
- Automatically controlled windows (1)
- Backdoors (1)
- Barcode (1)
- Batch processing (1)
- Bean (1)
- Benchmark testing; (1)
- Berliner Modell (1)
- Blind users (1)
- Boolean logic models (1)
- Bot Detection (1)
- Business Process (1)
- Business process intelligence (1)
- CP-Logic (1)
- Campus (1)
- Chipkarten (1)
- Circuit faults (1)
- Clock tree (1)
- Cloud (1)
- Cluster Computing (1)
- Cluster computing (1)
- Code generation (1)
- Codierungstheorie (1)
- Coherent phonons (1)
- Combinatorial multi-objective optimization (1)
- Complex optimization (1)
- Complexity (1)
- Computational complexity (1)
- Computational grid (1)
- Computer security (1)
- Computing with DNA (1)
- Conformant Planning (1)
- Conrad Hal Waddington (1)
- Constraint satisfaction (1)
- Context awareness (1)
- Contextualized learning (1)
- Continuous Testing (1)
- Continuous Versioning (1)
- Convolution (1)
- Course timetabling (1)
- Customer ownership (1)
- D-galactosamine (1)
- DMR (1)
- DNA hairpin formation (1)
- DRMAA (1)
- DRMS (1)
- Data federation (1)
- Database (1)
- Deal of the Day (1)
- Debugging (1)
- Decidability (1)
- Declarative Problem Solving (1)
- Declare (1)
- Deep Learning (1)
- Deep learning (1)
- Denotational semantics (1)
- Design (1)
- Design for testability (DFT) (1)
- Deutschland (1)
- Digital image analysis (1)
- Digitale Medien (1)
- Digitalization (1)
- Dose rate (1)
- Double cell upsets (DCUs) (1)
- Dynamical X-ray theory (1)
- E-teaching (1)
- EDC (1)
- EEG (1)
- Edge Computing (1)
- Educational game (1)
- Educational timetabling (1)
- Encoding (1)
- Engines (1)
- Enterprise Architecture (1)
- Entity Linking (1)
- Epigenetic landscape (1)
- Epistemic Logic Programs (1)
- Erklärbarkeit (1)
- Evaluation (1)
- Evolution (1)
- Experimentation (1)
- Explainability (1)
- Explicit negation (1)
- Explore-first Programming (1)
- Extensibility (1)
- Extreme Model-Driven Development (1)
- FEDC (1)
- FPGA (1)
- Fault Localization (1)
- Fault tolerant systems (1)
- Feature extraction (1)
- Fehlererkennung (1)
- Field programmable gate arrays (1)
- Finite automata (1)
- Flip-flops (1)
- Forgetting (1)
- Framework (1)
- Freshmen (1)
- GERBIL (1)
- Geschäftsprozess (1)
- Gesture input (1)
- Graph Convolutional Neural Networks (1)
- Graph Embedding (1)
- Green computing (1)
- Grounded theory (1)
- Grounding Theory (1)
- H.264 (1)
- Hairpin completions (1)
- Hairpin reductions (1)
- Hardware accelerator (1)
- Heat diffusion (1)
- Hierarchically configurable mask register (1)
- Histograms (1)
- Hochschul-Apps (1)
- Hochschul-Cloud (1)
- Hochschullehre (1)
- Https traffic (1)
- Human Factors (1)
- Hurricane Sandy (1)
- Hybrid App (1)
- IaaS (1)
- Identifiers (1)
- Image and video stylization (1)
- Image resolution (1)
- Imperative calculi (1)
- Improving classroom (1)
- Incoherent phonons (1)
- Incremental answer set programming (1)
- Industrie 4.0 (1)
- Industry 4.0 (1)
- Inference (1)
- Information federation (1)
- Information integration (1)
- Information retrieval (1)
- Information security (1)
- Insurance industry (1)
- Integrated circuit modeling (1)
- Interpretability (1)
- Interpretierbarkeit (1)
- Job monitoring (1)
- Job submission (1)
- Kernel (1)
- Kernelization (1)
- Key input (1)
- Knowledge (1)
- Knowledge Management (1)
- Knowledge representation (1)
- Kommunismus (1)
- Künstliche Neuronale Netzwerke (1)
- L systems (1)
- LBA problem (1)
- Landmark visibility (1)
- Learning (1)
- Lehre (1)
- Lernen (1)
- Literature mining (1)
- Liver neoplasms (1)
- Load Balancing (1)
- Localization (1)
- Location awareness (1)
- Logic programming (1)
- Loss (1)
- Low Latency (1)
- Loyalty (1)
- M2M (1)
- Markov processes (1)
- Masking of X-values (1)
- Massenklausuren (1)
- Media in education (1)
- Meta-Programming (1)
- Metric learning (1)
- Minimal perturbation problems (1)
- Mobile Campus Application (1)
- Mobile Learning (1)
- Mobile application (1)
- Mobile devices (1)
- Mobile learning (1)
- Model checking (1)
- Modellierung (1)
- Modelling (1)
- Multi-objective optimization (1)
- Multi-sided platforms (1)
- Multimedia (1)
- Multiple interpretation scheme (1)
- N-temperature model (1)
- NFC (1)
- Nash equilibrium (1)
- Natural language processing (1)
- Natural ventilation (1)
- Network (1)
- Network security (1)
- Neural networks (1)
- Non-Monotonic (1)
- Nonmonotonic reasoning (1)
- OBI (1)
- OCCI (1)
- OSSE (1)
- Open Badge Infrastructure (1)
- Open Badges (1)
- Operation problem (1)
- Optimierung (1)
- Parallel SAT solving (1)
- Parallel job execution time estimation (1)
- Particle detector (1)
- Pedagogical issues (1)
- Pedestrian navigation (1)
- Peer-Review (1)
- Performance Evaluation (1)
- Personal Learning Environment (1)
- Personalization (1)
- Pervasive computing (1)
- Pervasive game (1)
- Pervasive learning (1)
- Phantoms (1)
- Planar tactile display (1)
- Plant identification (1)
- Polarization (1)
- Privacy Protection (1)
- Process (1)
- Process Management (1)
- Process model analysis (1)
- Product lifecycle management (1)
- Professionalisierung (1)
- Programming (1)
- Programming by optimization (1)
- Prototyping (1)
- Prozess (1)
- Prozessmanagement (1)
- Prüfungsoptimierung (1)
- RADFET (1)
- RADFETs (1)
- REST (1)
- RSA triangle (1)
- Radiation hardness (1)
- Random access memory (1)
- Ranking (1)
- Reasoning (1)
- Reflexion (1)
- Region of Interest (1)
- Reproducibility of results (1)
- Reversibility (1)
- SAMR (1)
- SET pulsewidth distribution (1)
- SOA (1)
- SWOT (1)
- SaaSAbstract (1)
- Scalability (1)
- Scale-invariant feature transform (SIFT) (1)
- Scientific images (1)
- Screen reader (1)
- Seamless learning (1)
- Search problems (1)
- Security (1)
- Self-adaptive MPSoC (1)
- Semantic Interoperability (1)
- Semantic data (1)
- Semantic web (1)
- Semilinearity property (1)
- Sequence embeddings (1)
- Service orientation (1)
- Service-oriented Architecture (1)
- Service-oriented architecture (1)
- Sharing (1)
- Signal processing (1)
- Signaling transduction networks (1)
- Simulation (1)
- Simulations (1)
- Single event effect (1)
- Single event upsets (1)
- Single-event transient (SET) (1)
- Social Media Analysis (1)
- Speicher (1)
- Splicing (1)
- Splicing processor (1)
- Static Analysis (1)
- Statistical relational learning (1)
- Stochastic relational process (1)
- Strong equivalence (1)
- Structural equation modeling (1)
- Studieneinstieg (1)
- Studienorganisation (1)
- Systems biology (1)
- Systems of parallel communicating (1)
- TMR (1)
- Teilnehmerzertifikate (1)
- Temporal Answer Set Solving (1)
- Theory formation (1)
- Thermoelasticity (1)
- Time series (1)
- Tomography (1)
- Tools (1)
- Tracking (1)
- Traffic data (1)
- Tree decomposition (1)
- Treewidth (1)
- Treewidth-aware reductions (1)
- Triple modular redundancy (TMR) (1)
- Tumor types (1)
- Turing machine (1)
- Type and effect systems (1)
- UAV imagery (1)
- UX (1)
- Ubiquitous learning (1)
- Ultrafast dynamics (1)
- Unary languages (1)
- Uniform Access Principle (1)
- University Service Bus (1)
- Usability testing (1)
- User submission pattern (1)
- User-centred design (1)
- VGG16 (1)
- Value network (1)
- Verification (1)
- Virtual reality (1)
- Visual metaphor (1)
- Webanwendung (1)
- Wireless Sensor Networks (1)
- Wissen (1)
- Wissensmanagement (1)
- Word embeddings (1)
- X-masking (1)
- X-ray computed (1)
- X-values (1)
- accepting grammars (1)
- action and change (1)
- activities (1)
- activity (1)
- acute liver failure (1)
- acyclicity properties (1)
- adversarial classification (1)
- algorithm schedules (1)
- algorithms (1)
- analysis (1)
- annealing (1)
- anti-cancer drugs (1)
- anxiety (1)
- approximate model counting (1)
- architecture (1)
- argument mining (1)
- arousal (1)
- artistic rendering (1)
- asynchrounous design (1)
- authentication (1)
- autism (1)
- automata (1)
- automated driving (1)
- automated guided vehicle routing (1)
- automated planning (1)
- automatic feedback (1)
- behavioral (1)
- behavioral abstraction (1)
- belief merging (1)
- belief revision (1)
- benchmark (1)
- bibliometric analysis (1)
- block representation (1)
- bootstrapping (1)
- brain-computer interface (1)
- bundled data (1)
- camera sensor (1)
- car assembly operations (1)
- cellular automata (1)
- circuit Faults (1)
- citation analysis (1)
- click controller (1)
- clocks (1)
- co-citation analysis (1)
- co-occurrence analysis (1)
- code generation (1)
- coherence relation (1)
- collaborative learning (1)
- combinatorial optimization problems (1)
- combined task and motion planning (1)
- common spatial patterns (1)
- competition (1)
- complex networks (1)
- compliance (1)
- computer science education (1)
- computer vision (1)
- concession (1)
- concurrent checking (1)
- conductive argument (1)
- connective (1)
- connectivity (1)
- consistency (1)
- consistency checking (1)
- consistency measures (1)
- context-free grammar (1)
- context-sensitive (1)
- continuous (1)
- contrast (1)
- controlled vocabularies (1)
- corpus analysis (1)
- correlated errors (1)
- course timetabling (1)
- craters (1)
- crop (1)
- decidability questions (1)
- declarative problem solving (1)
- deep learning (1)
- deep neural networks (1)
- deep residual networks (1)
- degree of non-context-freeness (1)
- degree of non-regularity (1)
- degree of non-regulation (1)
- depression (1)
- design flow (1)
- determinism (1)
- detrending (1)
- developmental systems (1)
- diagnosis (1)
- domain-specific APIs (1)
- drug discovery (1)
- drug-sensitivity prediction (1)
- dynamic service binding (1)
- eLectures (1)
- economic ripples (1)
- edge computing (1)
- educational timetabling (1)
- embedded systems (1)
- emission factor (1)
- endothelin (1)
- endothelin-converting enzyme (1)
- ensemble kalman filter (1)
- ensemble methods (1)
- error propagation (1)
- evaluation (1)
- event-related desynchronization (1)
- evolution (1)
- external ambiguity (1)
- extreme weather (1)
- face tracking (1)
- facial expression (1)
- fading (1)
- fault tolerance (1)
- field-programmable gate array (1)
- finite model computation (1)
- finite state sequential transducers (1)
- firmware update (1)
- formal (1)
- formal argumentation systems (1)
- functions (1)
- gait (1)
- gap-filling (1)
- geovisualization (1)
- gradient boosting (1)
- grammar (1)
- graph analysis (1)
- greenhouse gas (1)
- hardware accelerator (1)
- hardware architecture (1)
- higher education (1)
- hybrid solving (1)
- ice harboring (1)
- image classification (1)
- image processing (1)
- image recognition (1)
- imaging (1)
- impacts (1)
- incremental SVM (1)
- informal logic (1)
- information flow control (1)
- internal ambiguity (1)
- intrusion detection (1)
- irradiation (1)
- joint lab (1)
- kidney cancer (1)
- knowledge representation and reasoning (1)
- latches (1)
- leftmost derivations (1)
- lesson planning (1)
- lesson preparation (1)
- linear programming (1)
- locomotion (1)
- logic programming (1)
- logic programming methodology and applications (1)
- logic-based modeling (1)
- loop formulas (1)
- loose programming (1)
- loss propagation (1)
- lunar exploration (1)
- machine learning (1)
- machine learning algorithms (1)
- manipulation planning (1)
- measure development (1)
- media (1)
- metabolic network (1)
- metabolism (1)
- metabolomics (1)
- metadata (1)
- metastasis (1)
- mobile Applikationen (1)
- natural disasters (1)
- natural language generation (1)
- neighborhood (1)
- neural networks (1)
- neutral endopeptidase (1)
- nonphotorealistic rendering (NPR) (1)
- o-ambiguity (1)
- on-farm evaluation (1)
- oneM2M Ontology (1)
- pMOS radiation dosimeter (1)
- parallel processing (1)
- parallel rewriting (1)
- parity aggregate operator (1)
- parsing (1)
- pdf forms (1)
- perception (1)
- perception differences (1)
- physical computing (1)
- planning (1)
- plug-ins (1)
- policy evaluation (1)
- portfolio-based solving (1)
- predictive models (1)
- premise acceptability (1)
- process model alignment (1)
- process modeling (1)
- program encodings (1)
- programmed grammars (1)
- projection (1)
- proof complexity (1)
- pruritus (1)
- pulse stretching inverters (1)
- quality of life (1)
- quantum (1)
- random forest (1)
- real arguments (1)
- real-time (1)
- real-time mapping (1)
- real-walking (1)
- reference (1)
- referential effectiveness (1)
- regression (1)
- regular language (1)
- relevance (1)
- reliability analysis (1)
- resources (1)
- restricted parallelism (1)
- risk analysis (1)
- safety (1)
- satisfiability (1)
- selective fault tolerance (1)
- self-adaptive multiprocessing system (1)
- self-checking (1)
- semantic web (1)
- sensitivity (1)
- simplicity (1)
- single event upset (1)
- single event upsets (1)
- single-event transient (1)
- single-trial-analysis (1)
- site-specific weed management (1)
- sleep quality (1)
- smart farming (1)
- smartphone (1)
- soft errors (1)
- solar particle event (1)
- space missions (1)
- stable model semantics (1)
- state complexity (1)
- static analysis (1)
- static prediction games (1)
- strong equivalence (1)
- sufficiency (1)
- suicidal ideations (1)
- supply chains (1)
- support system (1)
- support vector machines (1)
- tableau calculi (1)
- teacher training (1)
- teaching (1)
- technical notes and rapid communications (1)
- tele-teaching (1)
- test response compaction (1)
- theory (1)
- theory of computation (1)
- timing (1)
- tools (1)
- transient Faults (1)
- transient analysis (1)
- triangulated irregular networks (1)
- triple modular redundancy (1)
- unfounded sets (1)
- unidirektionale Fehler (1)
- user experience (1)
- video annotation (1)
- virtual mobility (1)
- wheat crops (1)
- work productivity (1)
- xAPI (1)
- yellow rust (1)
Institute
- Institut für Informatik und Computational Science (866)
- Institut für Physik und Astronomie (2)
- Department Erziehungswissenschaft (1)
- Department Linguistik (1)
- Extern (1)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (1)
- Institut für Biochemie und Biologie (1)
- Universitätsbibliothek (1)
- Zentrum für Qualitätsentwicklung in Lehre und Studium (ZfQ) (1)
Researchers and developers worldwide have put their efforts into the design, development and use of information and communication technology to support teaching and learning. This research is driven by pedagogical as well as technological disciplines. The most challenging ideas are currently found in the application of mobile, ubiquitous, pervasive, contextualized and seamless technologies for education, which we shall refer to as pervasive education. This article provides a comprehensive overview of the existing work in this field and categorizes it with respect to educational settings. Using this approach, best practice solutions for certain educational settings and open questions for pervasive education are highlighted in order to inspire interested developers and educators. The work is assigned to different fields, identified by the main pervasive technologies used and the educational settings. Based on these assignments we identify areas within pervasive education that are currently disregarded or deemed challenging so that further research and development in these fields are stimulated in a trans-disciplinary approach. (C) 2013 Elsevier B.V. All rights reserved.
We present a prototype of an integrated reasoning environment for educational purposes. The presented tool is a fragment of a proof assistant and automated theorem prover. We describe the existing and planned functionality of the theorem prover and especially the functionality of the educational fragment. This currently supports working with terms of the untyped lambda calculus and addresses both undergraduate students and researchers. We show how the tool can be used to support the students' understanding of functional programming and discuss general problems related to the process of building theorem proving software that aims at supporting both research and education.
We discuss the relaxation of a class of nonlinear elliptic Cauchy problems with data on a piece S of the boundary surface by means of a variational approach known in the optimal control literature as "equation error method". By the Cauchy problem is meant any boundary value problem for an unknown function y in a domain X with the property that the data on S, if combined with the differential equations in X, allow one to determine all derivatives of y on S by means of functional equations. In the case of real analytic data of the Cauchy problem, the existence of a local solution near S is guaranteed by the Cauchy-Kovalevskaya theorem. We also admit overdetermined elliptic systems, in which case the set of those Cauchy data on S for which the Cauchy problem is solvable is very "thin". For this reason we discuss a variational setting of the Cauchy problem which always possesses a generalised solution.
The ongoing digitalization leads to a need of continuous change of ICT (Information and Communi-cation Technology) in all university domains and therefore affects all stakeholders in this arena. More and more ICT components, systems and tools occur and have to be integrated into the existing processes and infrastructure of the institutions. These tasks include the transfer of resources and information across multiple ICT systems. By using so-called virtual environments for domains of re-search, education, learning and work, the performance of daily tasks can be aided. Based on a user requirement analysis different short- and long-term objectives were identified and are tackled now in the context of a federal research project. In order to be prepared for the ongoing digitalization, new systems have to be provided. Both, a service-oriented infrastructure and a related web-based virtual learning environment constitute the platform Campus.UP and creates the necessary basis to be ready for future challenges. The current focus lies on e-portfolio work, hence we will present a related focus group evaluation. The results indicate a tremendous need to extend the possibilities of sharing resources across system boundaries, in order to enable a comfortable participation of exter-nal cooperating parties and to clarify the focus of each connected system. The introduction of such an infrastructure implies far-reaching changes for traditional data centers. Therefore, the challenges and risks of faculty conducting innovation projects for the ICT organization are taken as a starting point to stimulate a discussion, how data centers can utilize projects to be ready for the future needs. We are confident that Campus.UP will provide the basis for ensuring the persistent transfer of innovation to the ICT organization and thus will contribute to tackle the future challenges of digitalization.
This paper describes the implementation of a workflow model for service-oriented computing of potential areas for wind turbines in jABC. By implementing a re-executable model the manual effort of a multi-criteria site analysis can be reduced. The aim is to determine the shift of typical geoprocessing tools of geographic information systems (GIS) from the desktop to the web. The analysis is based on a vector data set and mainly uses web services of the “Center for Spatial Information Science and Systems” (CSISS). This paper discusses effort, benefits and problems associated with the use of the web services.
Information integration across company borders becomes increasingly important for the success of product lifecycle management in industry and complex supply chains. Semantic technologies are about to play a crucial role in this integrative process. However, cross-company data exchange requires mechanisms to enable fine-grained access control definition and enforcement, preventing unauthorized leakage of confidential data across company borders. Currently available semantic repositories are not sufficiently equipped to satisfy this important requirement. This paper presents an infrastructure for controlled sharing of semantic data between cooperating business partners. First, we motivate the need for access control in semantic data federations by a case study in the industrial service sector. Furthermore, we present an architecture for controlling access to semantic repositories that is based on our newly developed SemForce security service. Finally, we show the practical feasibility of this architecture by an implementation and several performance experiments.
Evaluating the quality of ranking functions is a core task in web search and other information retrieval domains. Because query distributions and item relevance change over time, ranking models often cannot be evaluated accurately on held-out training data. Instead, considerable effort is spent on manually labeling the relevance of query results for test queries in order to track ranking performance. We address the problem of estimating ranking performance as accurately as possible on a fixed labeling budget. Estimates are based on a set of most informative test queries selected by an active sampling distribution. Query labeling costs depend on the number of result items as well as item-specific attributes such as document length. We derive cost-optimal sampling distributions for the commonly used performance measures Discounted Cumulative Gain and Expected Reciprocal Rank. Experiments on web search engine data illustrate significant reductions in labeling costs.
Owing to the loose coupling between replicas, the replica-exchange (RE) class of algorithms should be able to benefit greatly from using as many resources as available. However, the ability to effectively use multiple distributed resources to reduce the time to completion remains a challenge at many levels. Additionally, an implementation of a pleasingly distributed algorithm such as replica-exchange, which is independent of infrastructural details, does not exist. This paper proposes an extensible and scalable framework based on Simple API for Grid Applications that provides a general-purpose, opportunistic mechanism to effectively use multiple resources in an infrastructure-independent way. By analysing the requirements of the RE algorithm and the challenges of implementing it on real production systems, we propose a new abstraction (BIGJOB), which forms the basis of the adaptive redistribution and effective scheduling of replicas.
The introduction of columnar in-memory databases, along with hardware evolution, has made the execution of transactional and analytical enterprise application workloads on a single system both feasible and viable. Yet, we argue that executing analytical aggregate queries directly on the transactional data can decrease the overall system performance. Despite the aggregation capabilities of columnar in-memory databases, the direct access to records of a materialized aggregate is always more efficient than aggregating on the fly. The traditional approach to materialized aggregates, however, introduces significant overhead in terms of materialized view selection, maintenance, and exploitation. When this overhead is handled by the application, it increases the application complexity, and can slow down the transactional throughput of inserts, updates, and deletes.
In this thesis, we motivate, propose, and evaluate the aggregate cache, a materialized aggregate engine in the main-delta architecture of a columnar in-memory database that provides efficient means to handle costly aggregate queries of enterprise applications. For our design, we leverage the specifics of the main-delta architecture that separates a table into a main and delta partition. The central concept is to only cache the partial aggregate query result as defined on the main partition of a table, because the main partition is relatively stable as records are only inserted into the delta partition. We contribute by proposing incremental aggregate maintenance and query compensation techniques for mixed workloads of enterprise applications. In addition, we introduce aggregate profit metrics that increase the likelihood of persisting the most profitable aggregates in the aggregate cache.
Query compensation and maintenance of materialized aggregates based on joins of multiple tables is expensive due to the partitioned tables in the main-delta architecture. Our analysis of enterprise applications has revealed several data schema and workload patterns. This includes the observation that transactional data is persisted in header and item tables, whereas in many cases, the insertion of related header and item records is executed in a single database transaction. We contribute by proposing an approach to transport these application object semantics to the database system and optimize the query processing using the aggregate cache by applying partition pruning and predicate pushdown techniques.
For the experimental evaluation, we propose the FICO benchmark that is based on data from a productive ERP system with extracted mixed workloads. Our evaluation reveals that the aggregate cache can accelerate the execution of aggregate queries up to a factor of 60 whereas the speedup highly depends on the number of aggregated records in the main and delta partitions. In mixed workloads, the proposed aggregate maintenance and query compensation techniques perform up to an order of magnitude better than traditional materialized aggregate maintenance approaches. The introduced aggregate profit metrics outperform existing costbased metrics by up to 20%. Lastly, the join pruning and predicate pushdown techniques can accelerate query execution in the aggregate cache in the presence of multiple partitioned tables by up to an order of magnitude.
A multiple interpretation scheme is an ordered sequence of morphisms. The ordered multiple interpretation of a word is obtained by concatenating the images of that word in the given order of morphisms. The arbitrary multiple interpretation of a word is the semigroup generated by the images of that word. These interpretations are naturally extended to languages. Four types of ambiguity of multiple interpretation schemata on a language are defined: o-ambiguity, internal ambiguity, weakly external ambiguity and strongly external ambiguity. We investigate the problem of deciding whether a multiple interpretation scheme is ambiguous on regular languages.
Institutions are facing the challenge to integrate legacy systems with steadily growing new ones, using different technologies and interaction patterns. With the demand of offering the best potential of all systems, several not matching systems including their functions have to be aggregated and offered in a useable way. This paper presents an adaptive, generalizable and self-organized Personal Learning Environment (PLE) framework with the potential to integrate several heterogeneous services using a service-oriented architecture. First, a general overview over the field is given, followed by the description of the core components of the PLE framework. A prototypical implementation is presented. Finally, it’s shown how the PLE framework can be dynamically adapted to a changing system environment, reflecting experiences from first user studies.
An asymptotic analysis and improvement of AdaBoost in the binary classification case (in Japanese)
(2000)
Algorithm selection (AS) techniques - which involve choosing from a set of algorithms the one expected to solve a given problem instance most efficiently - have substantially improved the state of the art in solving many prominent AI problems, such as SAT, CSP, ASP, MAXSAT and QBF. Although several AS procedures have been introduced, not too surprisingly, none of them dominates all others across all AS scenarios. Furthermore, these procedures have parameters whose optimal values vary across AS scenarios. This holds specifically for the machine learning techniques that form the core of current AS procedures, and for their hyperparameters. Therefore, to successfully apply AS to new problems, algorithms and benchmark sets, two questions need to be answered: (i) how to select an AS approach and (ii) how to set its parameters effectively. We address both of these problems simultaneously by using automated algorithm configuration. Specifically, we demonstrate that we can automatically configure claspfolio 2, which implements a large variety of different AS approaches and their respective parameters in a single, highly-parameterized algorithm framework. Our approach, dubbed AutoFolio, allows researchers and practitioners across a broad range of applications to exploit the combined power of many different AS methods. We demonstrate AutoFolio can significantly improve the performance of claspfolio 2 on 8 out of the 13 scenarios from the Algorithm Selection Library, leads to new state-of-the-art algorithm selectors for 7 of these scenarios, and matches state-of-the-art performance (statistically) on all other scenarios. Compared to the best single algorithm for each AS scenario, AutoFolio achieves average speedup factors between 1.3 and 15.4.
An Extended Query language for action languages (and its application to aggregates and preferences)
(2006)
We address the problem of Finite Model Computation (FMC) of first-order theories and show that FMC can efficiently and transparently be solved by taking advantage of a recent extension of Answer Set Programming (ASP), called incremental Answer Set Programming (iASP). The idea is to use the incremental parameter in iASP programs to account for the domain size of a model. The FMC problem is then successively addressed for increasing domain sizes until an answer set, representing a finite model of the original first-order theory, is found. We implemented a system based on the iASP solver iClingo and demonstrate its competitiveness by showing that it slightly outperforms the winner of the FNT division of CADE's 2009 Automated Theorem Proving (ATP) competition on the respective benchmark collection.