@book{BauckmannLeserNaumann2010, author = {Bauckmann, Jana and Leser, Ulf and Naumann, Felix}, title = {Efficient and exact computation of inclusion dependencies for data integration}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, isbn = {978-3-86956-048-9}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41396}, publisher = {Universit{\"a}t Potsdam}, pages = {36}, year = {2010}, abstract = {Data obtained from foreign data sources often come with only superficial structural information, such as relation names and attribute names. Other types of metadata that are important for effective integration and meaningful querying of such data sets are missing. In particular, relationships among attributes, such as foreign keys, are crucial metadata for understanding the structure of an unknown database. The discovery of such relationships is difficult, because in principle for each pair of attributes in the database each pair of data values must be compared. A precondition for a foreign key is an inclusion dependency (IND) between the key and the foreign key attributes. We present with Spider an algorithm that efficiently finds all INDs in a given relational database. It leverages the sorting facilities of DBMS but performs the actual comparisons outside of the database to save computation. Spider analyzes very large databases up to an order of magnitude faster than previous approaches. We also evaluate in detail the effectiveness of several heuristics to reduce the number of necessary comparisons. Furthermore, we generalize Spider to find composite INDs covering multiple attributes, and partial INDs, which are true INDs for all but a certain number of values. This last type is particularly relevant when integrating dirty data as is often the case in the life sciences domain - our driving motivation.}, language = {en} } @inproceedings{GoltzPieth2010, author = {Goltz, Hans-Joachim and Pieth, Norbert}, title = {A tool for generating partition schedules of multiprocessor systems}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41556}, year = {2010}, abstract = {A deterministic cycle scheduling of partitions at the operating system level is supposed for a multiprocessor system. In this paper, we propose a tool for generating such schedules. We use constraint based programming and develop methods and concepts for a combined interactive and automatic partition scheduling system. This paper is also devoted to basic methods and techniques for modeling and solving this partition scheduling problem. Initial application of our partition scheduling tool has proved successful and demonstrated the suitability of the methods used.}, language = {en} } @inproceedings{BetzRaiserFruehwirth2010, author = {Betz, Hariolf and Raiser, Frank and Fr{\"u}hwirth, Thom}, title = {Persistent constraints in constraint handling rules}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41547}, year = {2010}, abstract = {In the most abstract definition of its operational semantics, the declarative and concurrent programming language CHR is trivially non-terminating for a significant class of programs. Common refinements of this definition, in closing the gap to real-world implementations, compromise on declarativity and/or concurrency. Building on recent work and the notion of persistent constraints, we introduce an operational semantics avoiding trivial non-termination without compromising on its essential features.}, language = {en} } @inproceedings{AbdennadherIsmailKhoury2010, author = {Abdennadher, Slim and Ismail, Haythem and Khoury, Frederick}, title = {Transforming imperative algorithms to constraint handling rules}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41533}, year = {2010}, abstract = {Different properties of programs, implemented in Constraint Handling Rules (CHR), have already been investigated. Proving these properties in CHR is fairly simpler than proving them in any type of imperative programming language, which triggered the proposal of a methodology to map imperative programs into equivalent CHR. The equivalence of both programs implies that if a property is satisfied for one, then it is satisfied for the other. The mapping methodology could be put to other beneficial uses. One such use is the automatic generation of global constraints, at an attempt to demonstrate the benefits of having a rule-based implementation for constraint solvers.}, language = {en} } @inproceedings{Brass2010, author = {Brass, Stefan}, title = {Range restriction for general formulas}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41521}, year = {2010}, abstract = {Deductive databases need general formulas in rule bodies, not only conjuctions of literals. This is well known since the work of Lloyd and Topor about extended logic programming. Of course, formulas must be restricted in such a way that they can be effectively evaluated in finite time, and produce only a finite number of new tuples (in each iteration of the TP-operator: the fixpoint can still be infinite). It is also necessary to respect binding restrictions of built-in predicates: many of these predicates can be executed only when certain arguments are ground. Whereas for standard logic programming rules, questions of safety, allowedness, and range-restriction are relatively easy and well understood, the situation for general formulas is a bit more complicated. We give a syntactic analysis of formulas that guarantees the necessary properties.}, language = {en} } @inproceedings{BandaGallagher2010, author = {Banda, Gourinath and Gallagher, John P.}, title = {Constraint-based abstraction of a model checker for infinite state systems}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41516}, year = {2010}, abstract = {Abstract interpretation-based model checking provides an approach to verifying properties of infinite-state systems. In practice, most previous work on abstract model checking is either restricted to verifying universal properties, or develops special techniques for temporal logics such as modal transition systems or other dual transition systems. By contrast we apply completely standard techniques for constructing abstract interpretations to the abstraction of a CTL semantic function, without restricting the kind of properties that can be verified. Furthermore we show that this leads directly to implementation of abstract model checking algorithms for abstract domains based on constraints, making use of an SMT solver.}, language = {en} } @inproceedings{HerreHummel2010, author = {Herre, Heinrich and Hummel, Axel}, title = {Stationary generated models of generalized logic programs}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41501}, year = {2010}, abstract = {The interest in extensions of the logic programming paradigm beyond the class of normal logic programs is motivated by the need of an adequate representation and processing of knowledge. One of the most difficult problems in this area is to find an adequate declarative semantics for logic programs. In the present paper a general preference criterion is proposed that selects the 'intended' partial models of generalized logic programs which is a conservative extension of the stationary semantics for normal logic programs of [Prz91]. The presented preference criterion defines a partial model of a generalized logic program as intended if it is generated by a stationary chain. It turns out that the stationary generated models coincide with the stationary models on the class of normal logic programs. The general wellfounded semantics of such a program is defined as the set-theoretical intersection of its stationary generated models. For normal logic programs the general wellfounded semantics equals the wellfounded semantics.}, language = {en} } @inproceedings{HerreHummel2010, author = {Herre, Heinrich and Hummel, Axel}, title = {A paraconsistent semantics for generalized logic programs}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41496}, year = {2010}, abstract = {We propose a paraconsistent declarative semantics of possibly inconsistent generalized logic programs which allows for arbitrary formulas in the body and in the head of a rule (i.e. does not depend on the presence of any specific connective, such as negation(-as-failure), nor on any specific syntax of rules). For consistent generalized logic programs this semantics coincides with the stable generated models introduced in [HW97], and for normal logic programs it yields the stable models in the sense of [GL88].}, language = {en} } @inproceedings{OetschSchwengererTompits2010, author = {Oetsch, Johannes and Schwengerer, Martin and Tompits, Hans}, title = {Kato: a plagiarism-detection tool for answer-set programs}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41485}, year = {2010}, abstract = {We present the tool Kato which is, to the best of our knowledge, the first tool for plagiarism detection that is directly tailored for answer-set programming (ASP). Kato aims at finding similarities between (segments of) logic programs to help detecting cases of plagiarism. Currently, the tool is realised for DLV programs but it is designed to handle various logic-programming syntax versions. We review basic features and the underlying methodology of the tool.}, language = {en} } @inproceedings{Cabalar2010, author = {Cabalar, Pedro}, title = {Existential quantifiers in the rule body}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41476}, year = {2010}, abstract = {In this paper we consider a simple syntactic extension of Answer Set Programming (ASP) for dealing with (nested) existential quantifiers and double negation in the rule bodies, in a close way to the recent proposal RASPL-1. The semantics for this extension just resorts to Equilibrium Logic (or, equivalently, to the General Theory of Stable Models), which provides a logic-programming interpretation for any arbitrary theory in the syntax of Predicate Calculus. We present a translation of this syntactic class into standard logic programs with variables (either disjunctive or normal, depending on the input rule heads), as those allowed by current ASP solvers. The translation relies on the introduction of auxiliary predicates and the main result shows that it preserves strong equivalence modulo the original signature.}, language = {en} } @inproceedings{GebserHinrichsSchaubetal.2010, author = {Gebser, Martin and Hinrichs, Henrik and Schaub, Torsten H. and Thiele, Sven}, title = {xpanda: a (simple) preprocessor for adding multi-valued propositions to ASP}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41466}, year = {2010}, abstract = {We introduce a simple approach extending the input language of Answer Set Programming (ASP) systems by multi-valued propositions. Our approach is implemented as a (prototypical) preprocessor translating logic programs with multi-valued propositions into logic programs with Boolean propositions only. Our translation is modular and heavily benefits from the expressive input language of ASP. The resulting approach, along with its implementation, allows for solving interesting constraint satisfaction problems in ASP, showing a good performance.}, language = {en} } @inproceedings{Seipel2010, author = {Seipel, Dietmar}, title = {Practical Applications of Extended Deductive Databases in DATALOG*}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41457}, year = {2010}, abstract = {A wide range of additional forward chaining applications could be realized with deductive databases, if their rule formalism, their immediate consequence operator, and their fixpoint iteration process would be more flexible. Deductive databases normally represent knowledge using stratified Datalog programs with default negation. But many practical applications of forward chaining require an extensible set of user-defined built-in predicates. Moreover, they often need function symbols for building complex data structures, and the stratified fixpoint iteration has to be extended by aggregation operations. We present an new language Datalog*, which extends Datalog by stratified meta-predicates (including default negation), function symbols, and user-defined built-in predicates, which are implemented and evaluated top-down in Prolog. All predicates are subject to the same backtracking mechanism. The bottom-up fixpoint iteration can aggregate the derived facts after each iteration based on user-defined Prolog predicates.}, language = {en} } @inproceedings{HanusKoschnicke2010, author = {Hanus, Michael and Koschnicke, Sven}, title = {An ER-based framework for declarative web programming}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41447}, year = {2010}, abstract = {We describe a framework to support the implementation of web-based systems to manipulate data stored in relational databases. Since the conceptual model of a relational database is often specified as an entity-relationship (ER) model, we propose to use the ER model to generate a complete implementation in the declarative programming language Curry. This implementation contains operations to create and manipulate entities of the data model, supports authentication, authorization, session handling, and the composition of individual operations to user processes. Furthermore and most important, the implementation ensures the consistency of the database w.r.t. the data dependencies specified in the ER model, i.e., updates initiated by the user cannot lead to an inconsistent state of the database. In order to generate a high-level declarative implementation that can be easily adapted to individual customer requirements, the framework exploits previous works on declarative database programming and web user interface construction in Curry.}, language = {en} } @inproceedings{Zhou2010, author = {Zhou, Neng-Fa}, title = {What I have learned from all these solver competitions}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41431}, year = {2010}, abstract = {In this talk, I would like to share my experiences gained from participating in four CSP solver competitions and the second ASP solver competition. In particular, I'll talk about how various programming techniques can make huge differences in solving some of the benchmark problems used in the competitions. These techniques include global constraints, table constraints, and problem-specific propagators and labeling strategies for selecting variables and values. I'll present these techniques with experimental results from B-Prolog and other CLP(FD) systems.}, language = {en} } @inproceedings{Schrijvers2010, author = {Schrijvers, Tom}, title = {Overview of the monadic constraint programming framework}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41411}, year = {2010}, abstract = {A constraint programming system combines two essential components: a constraint solver and a search engine. The constraint solver reasons about satisfiability of conjunctions of constraints, and the search engine controls the search for solutions by iteratively exploring a disjunctive search tree defined by the constraint program. The Monadic Constraint Programming framework gives a monadic definition of constraint programming where the solver is defined as a monad threaded through the monadic search tree. Search and search strategies can then be defined as firstclass objects that can themselves be built or extended by composable search transformers. Search transformers give a powerful and unifying approach to viewing search in constraint programming, and the resulting constraint programming system is first class and extremely flexible.}, language = {en} } @inproceedings{GeskeWolf2010, author = {Geske, Ulrich and Wolf, Armin}, title = {Preface}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41401}, year = {2010}, abstract = {The workshops on (constraint) logic programming (WLP) are the annual meeting of the Society of Logic Programming (GLP e.V.) and bring together researchers interested in logic programming, constraint programming, and related areas like databases, artificial intelligence and operations research. In this decade, previous workshops took place in Dresden (2008), W{\"u}rzburg (2007), Vienna (2006), Ulm (2005), Potsdam (2004), Dresden (2002), Kiel (2001), and W{\"u}rzburg (2000). Contributions to workshops deal with all theoretical, experimental, and application aspects of constraint programming (CP) and logic programming (LP), including foundations of constraint/ logic programming. Some of the special topics are constraint solving and optimization, extensions of functional logic programming, deductive databases, data mining, nonmonotonic reasoning, , interaction of CP/LP with other formalisms like agents, XML, JAVA, program analysis, program transformation, program verification, meta programming, parallelism and concurrency, answer set programming, implementation and software techniques (e.g., types, modularity, design patterns), applications (e.g., in production, environment, education, internet), constraint/logic programming for semantic web systems and applications, reasoning on the semantic web, data modelling for the web, semistructured data, and web query languages.}, language = {en} } @phdthesis{Bojdys2009, author = {Bojdys, Michael Janus}, title = {On new allotropes and nanostructures of carbon nitrides}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-41236}, school = {Universit{\"a}t Potsdam}, year = {2009}, abstract = {In the first section of the thesis graphitic carbon nitride was for the first time synthesised using the high-temperature condensation of dicyandiamide (DCDA) - a simple molecular precursor - in a eutectic salt melt of lithium chloride and potassium chloride. The extent of condensation, namely next to complete conversion of all reactive end groups, was verified by elemental microanalysis and vibrational spectroscopy. TEM- and SEM-measurements gave detailed insight into the well-defined morphology of these organic crystals, which are not based on 0D or 1D constituents like known molecular or short-chain polymeric crystals but on the packing motif of extended 2D frameworks. The proposed crystal structure of this g-C3N4 species was derived in analogy to graphite by means of extensive powder XRD studies, indexing and refinement. It is based on sheets of hexagonally arranged s-heptazine (C6N7) units that are held together by covalent bonds between C and N atoms. These sheets stack in a graphitic, staggered fashion adopting an AB-motif, as corroborated by powder X-ray diffractometry and high-resolution transmission electron microscopy. This study was contrasted with one of many popular - yet unsuccessful - approaches in the last 30 years of scientific literature to perform the condensation of an extended carbon nitride species through synthesis in the bulk. The second section expands the repertoire of available salt melts introducing the lithium bromide and potassium bromide eutectic as an excellent medium to obtain a new phase of graphitic carbon nitride. The combination of SEM, TEM, PXRD and electron diffraction reveals that the new graphitic carbon nitride phase stacks in an ABA' motif forming unprecedentedly large crystals. This section seizes the notion of the preceding chapter, that condensation in a eutectic salt melt is the key to obtain a high degree of conversion mainly through a solvatory effect. At the close of this chapter ionothermal synthesis is seen established as a powerful tool to overcome the inherent kinetic problems of solid state reactions such as incomplete polymerisation and condensation in the bulk especially when the temperature requirement of the reaction in question falls into the proverbial "no man's land" of classical solvents, i.e. above 250 to 300 °C. The following section puts the claim to the test, that the crystalline carbon nitrides obtained from a salt melt are indeed graphitic. A typical property of graphite - namely the accessibility of its interplanar space for guest molecules - is transferred to the graphitic carbon nitride system. Metallic potassium and graphitic carbon nitride are converted to give the potassium intercalation compound, K(C6N8)3 designated according to its stoichiometry and proposed crystal structure. Reaction of the intercalate with aqueous solvents triggers the exfoliation of the graphitic carbon nitride material and - for the first time - enables the access of singular (or multiple) carbon nitride sheets analogous to graphene as seen in the formation of sheets, bundles and scrolls of carbon nitride in TEM imaging. The thus exfoliated sheets form a stable, strongly fluorescent solution in aqueous media, which shows no sign in UV/Vis spectroscopy that the aromaticity of individual sheets was subject to degradation. The final section expands on the mechanism underlying the formation of graphitic carbon nitride by literally expanding the distance between the covalently linked heptazine units which constitute these materials. A close examination of all proposed reaction mechanisms to-date in the light of exhaustive DSC/MS experiments highlights the possibility that the heptazine unit can be formed from smaller molecules, even if some of the designated leaving groups (such as ammonia) are substituted by an element, R, which later on remains linked to the nascent heptazine. Furthermore, it is suggested that the key functional groups in the process are the triazine- (Tz) and the carbonitrile- (CN) group. On the basis of these assumptions, molecular precursors are tailored which encompass all necessary functional groups to form a central heptazine unit of threefold, planar symmetry and then still retain outward functionalities for self-propagated condensation in all three directions. Two model systems based on a para-aryl (ArCNTz) and para-biphenyl (BiPhCNTz) precursors are devised via a facile synthetic procedure and then condensed in an ionothermal process to yield the heptazine based frameworks, HBF-1 and HBF-2. Due to the structural motifs of their molecular precursors, individual sheets of HBF-1 and HBF-2 span cavities of 14.2 {\AA} and 23.0 {\AA} respectively which makes both materials attractive as potential organic zeolites. Crystallographic analysis confirms the formation of ABA' layered, graphitic systems, and the extent of condensation is confirmed as next-to-perfect by elemental analysis and vibrational spectroscopy.}, language = {en} } @book{WistSchaeferVogleretal.2010, author = {Wist, Dominic and Schaefer, Mark and Vogler, Walter and Wollowski, Ralf}, title = {STG decomposition : internal communication for SI implementability}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, isbn = {978-3-86956-037-3}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-40786}, publisher = {Universit{\"a}t Potsdam}, pages = {36}, year = {2010}, abstract = {STG decomposition is a promising approach to tackle the complexity problems arising in logic synthesis of speed independent circuits, a robust asynchronous (i.e. clockless) circuit type. Unfortunately, STG decomposition can result in components that in isolation have irreducible CSC conflicts. Generalising earlier work, it is shown how to resolve such conflicts by introducing internal communication between the components via structural techniques only.}, language = {en} } @phdthesis{Breitenbach2009, author = {Breitenbach, Sebastian Franz Martin}, title = {Changes in monsoonal precipitation and atmospheric circulation during the Holocene reconstructed from stalagmites from Northeastern India}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-37807}, school = {Universit{\"a}t Potsdam}, year = {2009}, abstract = {Recent years witnessed a vast advent of stalagmites as palaeoclimate archives. The multitude of geochemical and physical proxies and a promise of a precise and accurate age model greatly appeal to palaeoclimatologists. Although substantial progress was made in speleothem-based palaeoclimate research and despite high-resolution records from low-latitudinal regions, proving that palaeo-environmental changes can be archived on sub-annual to millennial time scales our comprehension of climate dynamics is still fragmentary. This is in particular true for the summer monsoon system on the Indian subcontinent. The Indian summer monsoon (ISM) is an integral part of the intertropical convergence zone (ITCZ). As this rainfall belt migrates northward during boreal summer, it brings monsoonal rainfall. ISM strength depends however on a variety of factors, including snow cover in Central Asia and oceanic conditions in the Indic and Pacific. Presently, many of the factors influencing the ISM are known, though their exact forcing mechanism and mutual relations remain ambiguous. Attempts to make an accurate prediction of rainfall intensity and frequency and drought recurrence, which is extremely important for South Asian countries, resemble a puzzle game; all interaction need to fall into the right place to obtain a complete picture. My thesis aims to create a faithful picture of climate change in India, covering the last 11,000 ka. NE India represents a key region for the Bay of Bengal (BoB) branch of the ISM, as it is here where the monsoon splits into a northwestward and a northeastward directed arm. The Meghalaya Plateau is the first barrier for northward moving air masses and receives excessive summer rainfall, while the winter season is very dry. The proximity of Meghalaya to the Tibetan Plateau on the one hand and the BoB on the other hand make the study area a key location for investigating the interaction between different forcings that governs the ISM. A basis for the interpretation of palaeoclimate records, and a first important outcome of my thesis is a conceptual model which explains the observed pattern of seasonal changes in stable isotopes (d18O and d2H) in rainfall. I show that although in tropical and subtropical regions the amount effect is commonly called to explain strongly depleted isotope values during enhanced rainfall, alone it cannot account for observed rainwater isotope variability in Meghalaya. Monitoring of rainwater isotopes shows no expected negative correlation between precipitation amount and d18O of rainfall. In turn I find evidence that the runoff from high elevations carries an inherited isotopic signature into the BoB, where during the ISM season the freshwater builds a strongly depleted plume on top of the marine water. The vapor originating from this plume is likely to memorize' and transmit further very negative d18O values. The lack of data does not allow for quantication of this plume effect' on isotopes in rainfall over Meghalaya but I suggest that it varies on seasonal to millennial timescales, depending on the runoff amount and source characteristics. The focal point of my thesis is the extraction of climatic signals archived in stalagmites from NE India. High uranium concentration in the stalagmites ensured excellent age control required for successful high-resolution climate reconstructions. Stable isotope (d18O and d13C) and grey-scale data allow unprecedented insights into millennial to seasonal dynamics of the summer and winter monsoon in NE India. ISM strength (i. e. rainfall amount) is recorded in changes in d18Ostalagmites. The d13C signal, reflecting drip rate changes, renders a powerful proxy for dry season conditions, and shows similarities to temperature-related changes on the Tibetan Plateau. A sub-annual grey-scale profile supports a concept of lower drip rate and slower stalagmite growth during dry conditions. During the Holocene, ISM followed a millennial-scale decrease of insolation, with decadal to centennial failures resulting from atmospheric changes. The period of maximum rainfall and enhanced seasonality corresponds to the Holocene Thermal Optimum observed in Europe. After a phase of rather stable conditions, 4.5 kyr ago, the strengthening ENSO system dominated the ISM. Strong El Nino events weakened the ISM, especially when in concert with positive Indian Ocean dipole events. The strongest droughts of the last 11 kyr are recorded during the past 2 kyr. Using the advantage of a well-dated stalagmite record at hand I tested the application of laser ablation-inductively coupled plasma-mass spectrometry (LA-ICP-MS) to detect sub-annual to sub-decadal changes in element concentrations in stalagmites. The development of a large ablation cell allows for ablating sample slabs of up to 22 cm total length. Each analyzed element is a potential proxy for different climatic parameters. Combining my previous results with the LAICP- MS-generated data shows that element concentration depends not only on rainfall amount and associated leaching from the soil. Additional factors, like biological activity and hydrogeochemical conditions in the soil and vadose zone can eventually affect the element content in drip water and in stalagmites. I present a theoretical conceptual model for my study site to explain how climatic signals can be transmitted and archived in stalagmite carbonate. Further, I establish a first 1500 year long element record, reconstructing rainfall variability. Additionally, I hypothesize that volcanic eruptions, producing large amounts of sulfuric acid, can influence soil acidity and hence element mobilization.}, language = {en} } @book{SmirnovWeidlichMendlingetal.2009, author = {Smirnov, Sergey and Weidlich, Matthias and Mendling, Jan and Weske, Mathias}, title = {Action patterns in business process models}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, isbn = {978-3-86956-009-0}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-33586}, publisher = {Universit{\"a}t Potsdam}, pages = {19}, year = {2009}, abstract = {Business process management experiences a large uptake by the industry, and process models play an important role in the analysis and improvement of processes. While an increasing number of staff becomes involved in actual modeling practice, it is crucial to assure model quality and homogeneity along with providing suitable aids for creating models. In this paper we consider the problem of offering recommendations to the user during the act of modeling. Our key contribution is a concept for defining and identifying so-called action patterns - chunks of actions often appearing together in business processes. In particular, we specify action patterns and demonstrate how they can be identified from existing process model repositories using association rule mining techniques. Action patterns can then be used to suggest additional actions for a process model. Our approach is challenged by applying it to the collection of process models from the SAP Reference Model.}, language = {en} }