004 Datenverarbeitung; Informatik
Refine
Has Fulltext
- yes (55)
Year of publication
- 2010 (55) (remove)
Document Type
- Conference Proceeding (25)
- Article (13)
- Monograph/Edited Volume (10)
- Doctoral Thesis (5)
- Postprint (2)
Language
- English (39)
- German (15)
- Multiple languages (1)
Keywords
- middleware (4)
- Aspektorientierte Softwareentwicklung (2)
- Betriebssysteme (2)
- Constraint Solving (2)
- Deduction (2)
- Informationsextraktion (2)
- Logic Programming (2)
- Logics (2)
- Middleware (2)
- Optimization (2)
- Planing (2)
- Systemsoftware (2)
- Virtuelle Maschinen (2)
- operating systems (2)
- systems software (2)
- virtual machines (2)
- 3D city models (1)
- 3D geovisualisation (1)
- 3D-Geovisualisierung (1)
- 3D-Stadtmodelle (1)
- Asynchrone Schaltung (1)
- Asynchronous circuit (1)
- Ausbildung (1)
- Business Process Models (1)
- CSC (1)
- Coccinelle (1)
- Compliance (1)
- Computer Networks (1)
- Computernetzwerke (1)
- Controller-Resynthese (1)
- Crime mapping (1)
- Datenanalyse (1)
- Datenintegration (1)
- Didaktik (1)
- Dienstkomposition (1)
- Dienstplattform (1)
- Dynamic Type System (1)
- Dynamische Typ Systeme (1)
- Enterprise Search (1)
- Exploration (1)
- Fehlersuche (1)
- Forschungskolleg (1)
- GIS (1)
- Geschäftsprozessmodelle (1)
- Grammatikalische Inferenz (1)
- Graph-basiertes Ranking (1)
- Hasso Plattner Institute (1)
- Hasso-Plattner-Institut (1)
- Herodotos (1)
- High-Level Synthesis (1)
- History of pattern occurrences (1)
- Hochschuldidaktik (1)
- IPv4 (1)
- IPv6 (1)
- Informatik (1)
- Information Extraction (1)
- Infrastructure (1)
- Infrastruktur (1)
- Internet (1)
- Internet Protocol (1)
- Klausurtagung (1)
- Konferenz (1)
- Laufzeitanalyse (1)
- Linked Data (1)
- Metadatenentdeckung (1)
- Metadatenqualität (1)
- Multiprocessor (1)
- Multiprozessor (1)
- Muster (1)
- Musterabgleich (1)
- Network Politics (1)
- Netzpolitik (1)
- Next Generation Network (1)
- Object-Oriented Programming (1)
- Objekt-Orientiertes Programmieren (1)
- Parallel Programming (1)
- Parallele Datenverarbeitung (1)
- Parallelrechner (1)
- Pattern Matching (1)
- Patterns (1)
- Petrinetz (1)
- Ph.D. Retreat (1)
- Policy Enforcement (1)
- Programmierung (1)
- Prolog (1)
- Reconfigurable (1)
- Research School (1)
- SOA (1)
- STG decomposition (1)
- STG-Dekomposition (1)
- Schemaentdeckung (1)
- Service Creation (1)
- Service Delivery Platform (1)
- Service Provider (1)
- Service convergence (1)
- Service-oriented Systems Engineering (1)
- Signalflankengraph (SFG oder STG) (1)
- Smalltalk (1)
- Synthese (1)
- Telekommunikation (1)
- Temporal Logic (1)
- Temporallogik (1)
- VIL (1)
- Verletzung Auflösung (1)
- Verletzung Erklärung (1)
- Violation Explanation (1)
- Violation Resolution (1)
- Web-Anwendungen (1)
- Wikipedia (1)
- XM (1)
- aspect adapter (1)
- aspect oriented programming (1)
- aspect-oriented (1)
- aspects (1)
- aspectualization (1)
- bottom–up (1)
- bug tracking (1)
- built–in predicates (1)
- concurrency (1)
- conference (1)
- consistency (1)
- control resynthesis (1)
- crosscutting wrappers (1)
- data integration (1)
- data profiling (1)
- debugging (1)
- deductive databases (1)
- didactics (1)
- differential gene expression (1)
- distributed systems (1)
- dynamic reconfiguration (1)
- education (1)
- enterprise search (1)
- evolving systems (1)
- extensions of logic programs (1)
- forward / backward chaining (1)
- function symbols (1)
- gene expression matrix (1)
- generalized logic programs (1)
- geovirtual environments (1)
- geovirtuelle Umgebungen (1)
- grammar inference (1)
- graph-based ranking (1)
- heterogeneous tissue (1)
- higher education (1)
- homogeneous cell population (1)
- informatics (1)
- information extraction (1)
- invasive aspects (1)
- knowledge representation (1)
- konvergente Dienste (1)
- metadata discovery (1)
- metadata quality (1)
- microdissection (1)
- multi core data processing (1)
- multi-valued logic (1)
- paraconsistency (1)
- petri net (1)
- profiling (1)
- program analysis (1)
- programming (1)
- programming language (1)
- quantile normalization (1)
- raumbezogene Straftatenanalyse (1)
- reflection (1)
- rekonfigurierbar (1)
- reusable aspects (1)
- schema discovery (1)
- security policies (1)
- selection (1)
- semantic web services (1)
- semantics (1)
- service mediation (1)
- service-oriented (1)
- signal transition graph (1)
- smalltalk (1)
- speed independent (1)
- stratification (1)
- top– down (1)
- tuple spaces (1)
- views (1)
- web services (1)
- web-applications (1)
Companies develop process models to explicitly describe their business operations. In the same time, business operations, business processes, must adhere to various types of compliance requirements. Regulations, e.g., Sarbanes Oxley Act of 2002, internal policies, best practices are just a few sources of compliance requirements. In some cases, non-adherence to compliance requirements makes the organization subject to legal punishment. In other cases, non-adherence to compliance leads to loss of competitive advantage and thus loss of market share. Unlike the classical domain-independent behavioral correctness of business processes, compliance requirements are domain-specific. Moreover, compliance requirements change over time. New requirements might appear due to change in laws and adoption of new policies. Compliance requirements are offered or enforced by different entities that have different objectives behind these requirements. Finally, compliance requirements might affect different aspects of business processes, e.g., control flow and data flow. As a result, it is infeasible to hard-code compliance checks in tools. Rather, a repeatable process of modeling compliance rules and checking them against business processes automatically is needed. This thesis provides a formal approach to support process design-time compliance checking. Using visual patterns, it is possible to model compliance requirements concerning control flow, data flow and conditional flow rules. Each pattern is mapped into a temporal logic formula. The thesis addresses the problem of consistency checking among various compliance requirements, as they might stem from divergent sources. Also, the thesis contributes to automatically check compliance requirements against process models using model checking. We show that extra domain knowledge, other than expressed in compliance rules, is needed to reach correct decisions. In case of violations, we are able to provide a useful feedback to the user. The feedback is in the form of parts of the process model whose execution causes the violation. In some cases, our approach is capable of providing automated remedy of the violation.
We propose a paraconsistent declarative semantics of possibly inconsistent generalized logic programs which allows for arbitrary formulas in the body and in the head of a rule (i.e. does not depend on the presence of any specific connective, such as negation(-as-failure), nor on any specific syntax of rules). For consistent generalized logic programs this semantics coincides with the stable generated models introduced in [HW97], and for normal logic programs it yields the stable models in the sense of [GL88].
A deterministic cycle scheduling of partitions at the operating system level is supposed for a multiprocessor system. In this paper, we propose a tool for generating such schedules. We use constraint based programming and develop methods and concepts for a combined interactive and automatic partition scheduling system. This paper is also devoted to basic methods and techniques for modeling and solving this partition scheduling problem. Initial application of our partition scheduling tool has proved successful and demonstrated the suitability of the methods used.
We describe a framework to support the implementation of web-based systems to manipulate data stored in relational databases. Since the conceptual model of a relational database is often specified as an entity-relationship (ER) model, we propose to use the ER model to generate a complete implementation in the declarative programming language Curry. This implementation contains operations to create and manipulate entities of the data model, supports authentication, authorization, session handling, and the composition of individual operations to user processes. Furthermore and most important, the implementation ensures the consistency of the database w.r.t. the data dependencies specified in the ER model, i.e., updates initiated by the user cannot lead to an inconsistent state of the database. In order to generate a high-level declarative implementation that can be easily adapted to individual customer requirements, the framework exploits previous works on declarative database programming and web user interface construction in Curry.
In der letzten Jahren ist die Zahl der erfolgreichen Prüfungen von Studierenden im Informatikkurs des ersten Studienjahres für verschiedene Studiengänge an der Universität Óbuda stark gesunken. Dies betrifft Prüfungen in den Teilgebieten Rechnerarchitektur, Betrieb von Peripheriegeräten, Binäre Codierung und logische Operationen, Computerviren, Computernetze und das Internet, Steganographie und Kryptographie, Betriebsysteme. Mehr als der Hälfte der Studenten konnte die Prüfungen der ersten Semester nicht erfolgreich absolvieren. Die hier vorgelegte Analyse der Studienleistungen zielt darauf ab, Gründe für diese Entwicklung zu identifizieren, die Zahl der Abbrecher zu reduzieren und die Leistungen der Studenten zu verbessern. Die Analyse zeigt, dass die Studenten die erforderlichen Lehrmaterialen erst ein bis zwei Tage vor oder sogar erst am Tag der Klausuren vom Server downloaden, so dass sie nicht mehr hinreichend Zeit zum Lernen haben. Diese Tendenz zeigt sich bei allen Teilgebieten des Studiengangs. Ein Mangel an kontinuierlicher Mitarbeit scheint einer der Gründe für ein frühes Scheitern zu sein. Ferner zeigt sich die Notwendigkeit, dass bei den Lehrangeboten in Informatik auf eine kontinuierliche Kommunikation mit den Studierenden und Rückmeldung zu aktuellen Unterrichtsinhalten zu achten ist. Dies kann durch motivierende Maßnahmen zur Teilnahme an den Übungen oder durch kleine wöchentliche schriftliche Tests geschehen.
Projektmanagement-Kompetenzen werden von Unternehmen unterschiedlichster Branchen mit wachsender Priorität betrachtet und eingefordert. Als Beitrag zu einer kompetenzorientierten Ausbildung werden in diesem Paper interdisziplinäre Studienmodule als Bestandteil des Wirtschaftsinformatik-Studiums vorgestellt. Zielsetzung der Studienmodule ist die Befähigung der Studierenden, konkrete Projekte unter Nutzung von standardisierten Werkzeugen und Methoden nach dem IPMA-Standard planen und durchführen zu können.
This thesis presents methods for automated synthesis of flexible chip multiprocessor systems from parallel programs targeted at FPGAs to exploit both task-level parallelism and architecture customization. Automated synthesis is necessitated by the complexity of the design space. A detailed description of the design space is provided in order to determine which parameters should be modeled to facilitate automated synthesis by optimizing a cost function, the emphasis being placed on inclusive modeling of parameters from application, architectural and physical subspaces, as well as their joint coverage in order to avoid pre-constraining the design space. Given a parallel program and a set of an IP library, the automated synthesis problem is to simultaneously (i) select processors (ii) map and schedule tasks to them, and (iii) select one or several networks for inter-task communications such that design constraints and optimization objectives are met. The research objective in this thesis is to find a suitable model for automated synthesis, and to evaluate methods of using the model for architectural optimizations. Our contributions are a holistic approach for the design of such systems, corresponding models to facilitate automated synthesis, evaluation of optimization methods using state of the art integer linear and answer set programming, as well as the development of synthesis heuristics to solve runtime challenges.
Enforcing security policies to distributed systems is difficult, in particular, when a system contains untrusted components. We designed AspectKE*, a distributed AOP language based on a tuple space, to tackle this issue. In AspectKE*, aspects can enforce access control policies that depend on future behavior of running processes. One of the key language features is the predicates and functions that extract results of static program analysis, which are useful for defining security aspects that have to know about future behavior of a program. AspectKE* also provides a novel variable binding mechanism for pointcuts, so that pointcuts can uniformly specify join points based on both static and dynamic information about the program. Our implementation strategy performs fundamental static analysis at load-time, so as to retain runtime overheads minimal. We implemented a compiler for AspectKE*, and demonstrate usefulness of AspectKE* through a security aspect for a distributed chat system.
Background: For heterogeneous tissues, such as blood, measurements of gene expression are confounded by relative proportions of cell types involved. Conclusions have to rely on estimation of gene expression signals for homogeneous cell populations, e.g. by applying micro-dissection, fluorescence activated cell sorting, or in-silico deconfounding. We studied feasibility and validity of a non-negative matrix decomposition algorithm using experimental gene expression data for blood and sorted cells from the same donor samples. Our objective was to optimize the algorithm regarding detection of differentially expressed genes and to enable its use for classification in the difficult scenario of reversely regulated genes. This would be of importance for the identification of candidate biomarkers in heterogeneous tissues.
Results: Experimental data and simulation studies involving noise parameters estimated from these data revealed that for valid detection of differential gene expression, quantile normalization and use of non-log data are optimal. We demonstrate the feasibility of predicting proportions of constituting cell types from gene expression data of single samples, as a prerequisite for a deconfounding-based classification approach. Classification cross-validation errors with and without using deconfounding results are reported as well as sample-size dependencies. Implementation of the algorithm, simulation and analysis scripts are available.
Conclusions: The deconfounding algorithm without decorrelation using quantile normalization on non-log data is proposed for biomarkers that are difficult to detect, and for cases where confounding by varying proportions of cell types is the suspected reason. In this case, a deconfounding ranking approach can be used as a powerful alternative to, or complement of, other statistical learning approaches to define candidate biomarkers for molecular diagnosis and prediction in biomedicine, in realistically noisy conditions and with moderate sample sizes.
Business process management aims at capturing, understanding, and improving work in organizations. The central artifacts are process models, which serve different purposes. Detailed process models are used to analyze concrete working procedures, while high-level models show, for instance, handovers between departments. To provide different views on process models, business process model abstraction has emerged. While several approaches have been proposed, a number of abstraction use case that are both relevant for industry and scientifically challenging are yet to be addressed. In this paper we systematically develop, classify, and consolidate different use cases for business process model abstraction. The reported work is based on a study with BPM users in the health insurance sector and validated with a BPM consultancy company and a large BPM vendor. The identified fifteen abstraction use cases reflect the industry demand. The related work on business process model abstraction is evaluated against the use cases, which leads to a research agenda.