@book{MarienfeldSogomonyanOcheretnijetal.2005, author = {Marienfeld, Daniel and Sogomonyan, Egor S. and Ocheretnij, V. and G{\"o}ssel, Michael}, title = {Self-checking Output-duplicated Booth-2 Multiplier}, series = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, volume = {2005, 1}, journal = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, publisher = {Univ.}, address = {Potsdam}, issn = {0946-7580}, year = {2005}, language = {en} } @book{BoernerGoessel2005, author = {B{\"o}rner, Ferdinand and G{\"o}ssel, Michael}, title = {Grundlagen digitaler Systeme}, publisher = {Univ.-Verl.}, address = {Potsdam}, isbn = {978-3-937786-46-9}, pages = {95 S.}, year = {2005}, language = {de} } @book{KuropkaMeyer2005, author = {Kuropka, Dominik and Meyer, Harald}, title = {Survey on service composition}, series = {Technische Berichte des Hasso-Plattner-Instituts f{\"u}r Softwaresystemtechnik an der Universit{\"a}t Potsda}, volume = {10}, journal = {Technische Berichte des Hasso-Plattner-Instituts f{\"u}r Softwaresystemtechnik an der Universit{\"a}t Potsda}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, isbn = {3-937786-78-3}, issn = {1613-5652}, pages = {35 S.}, year = {2005}, language = {en} } @article{KonczakVogel2005, author = {Konczak, Kathrin and Vogel, Ralf}, title = {Abduction and Preferences in Linguistics}, year = {2005}, language = {en} } @article{CaludeJurgensen2005, author = {Calude, C. S. and Jurgensen, Helmut}, title = {Is complexity a source of incompleteness?}, issn = {0196-8858}, year = {2005}, abstract = {In this paper we prove Chaitin's "heuristic principle," the theorems of a finitely-specified theory cannot be significantly more complex than the theory itself, for an appropriate measure of complexity. We show that the measure is invariant under the change of the Godel numbering. For this measure, the theorems of a finitely-specified, sound, consistent theory strong enough to formalize arithmetic which is arithmetically sound (like Zermelo-Fraenkel set theory with choice or Peano Arithmetic) have bounded complexity, hence every sentence of the theory which is significantly more complex than the theory is unprovable. Previous results showing that incompleteness is not accidental, but ubiquitous are here reinforced in probabilistic terms: the probability that a true sentence of length n is provable in the theory tends to zero when n tends to infinity, while the probability that a sentence of length n is true is strictly positive. (c) 2004 Elsevier Inc. All rights reserved}, language = {en} } @article{Bordihn2005, author = {Bordihn, Henning}, title = {On the number of components in cooperating distributed grammar systems}, issn = {0304-3975}, year = {2005}, abstract = {It is proved that the number of components in context-free cooperating distributed (CD) grammar systems can be reduced to 3 when they are working in the so-called sf-mode of derivation, which is the cooperation protocol which has been considered first for CD grammar systems. In this derivation mode, a component continues the derivation until and unless there is a nonterminal in the sentential form which cannot be rewritten according to that component. Moreover, it is shown that CD grammar systems in sf-mode with only one component can generate only the context-free languages but they can generate non-context-free languages if two components are used. The sf-mode of derivation is compared with other well-known cooperation protocols with respect to the hierarchies induced by the number of components. (C) 2004 Elsevier B.V. All rights reserved}, language = {en} } @article{BeerenwinkelSingLengaueretal.2005, author = {Beerenwinkel, Niko and Sing, Tobias and Lengauer, Thomas and Rahnenfuhrer, Joerg and Roomp, Kirsten and Savenkov, Igor and Fischer, Roman and Hoffmann, Daniel and Selbig, Joachim and Korn, Klaus and Walter, Hauke and Berg, Thomas and Braun, Patrick and Faetkenheuer, Gerd and Oette, Mark and Rockstroh, Juergen and Kupfer, Bernd and Kaiser, Rolf and Daeumer, Martin}, title = {Computational methods for the design of effective therapies against drug resistant HIV strains}, year = {2005}, abstract = {The development of drug resistance is a major obstacle to successful treatment of HIV infection. The extraordinary replication dynamics of HIV facilitates its escape from selective pressure exerted by the human immune system and by combination drug therapy. We have developed several computational methods whose combined use can support the design of optimal antiretroviral therapies based on viral genomic data}, language = {en} } @article{BrzozowskiJuergensen2005, author = {Brzozowski, J. A. and J{\"u}rgensen, Helmut}, title = {Representation of semiautomata by canonical words and equivalences}, issn = {0129-0541}, year = {2005}, abstract = {We study a novel representation of semiautomata, which is motivated by the method of trace-assertion specifications of software modules. Each state of the semiautomaton is represented by an arbitrary word leading to that state, the canonical word. The transitions of the semiautomaton give rise to a right congruence, the state-equivalence, on the set of input words of the semiautomaton: two words are state-equivalent if and only if they lead to the same state. We present a simple algorithm for finding a set of generators for state-equivalence. Directly from this set of generators, we construct a confluent prefix-rewriting system which permits us to transform any word to its canonical representative. In general, the rewriting system may allow infinite derivations. To address this issue, we impose the condition of prefix-continuity on the set of canonical words. A set is prefix-continuous if, whenever a word w and a prefix u of w axe in the set, then all the prefixes of w longer than u are also in the set. Prefix-continuous sets include prefix-free and prefix-closed sets as special cases. We prove that the rewriting system is Noetherian if and only if the set of canonical words is prefix-continuous. Furthermore, if the set of canonical words is prefix- continuous, then the set of rewriting rules is irredundant. We show that each prefix-continuous canonical set corresponds to a spanning forest of the semiautomaton}, language = {en} } @article{BordihnHolzerKutrib2005, author = {Bordihn, Henning and Holzer, Markus and Kutrib, Martin}, title = {Unsolvability levels of operation problems for subclasses of context-free languages}, issn = {0129-0541}, year = {2005}, abstract = {We investigate the operation problem for linear and deterministic context-free languages: Fix an operation on formal languages. Given linear (deterministic, respectively) context-free languages, is the application of this operation to the given languages still a linear (deterministic, respectively) context-free language? Besides the classical operations, for which the linear and deterministic context-free languages are not closed, we also consider the recently introduced root and power operation. We show non-semidecidability, to be more precise, we show completeness for the second level of the arithmetic hierarchy for all of the aforementioned operations, except for the power operation, if the underlying alphabet contains at least two letters. The result for the power opera, tion solves an open problem stated in Theoret. Comput. Sci. 314 (2004) 445-449}, language = {en} } @article{BruggemeierDovifatKubisch2005, author = {Bruggemeier, M. and Dovifat, A. and Kubisch, D.}, title = {Micropolitical innovation arenas as a tool for analyzing innovation processes in the context of electronic government}, issn = {0937-6429}, year = {2005}, abstract = {E-Government requires technical and organizational innovation. Research has already shown that the respective innovation process is complex and contingent upon specific organizational structures. Managing such innovation processes successfully is difficult. Drawing on assumptions of micropolitical behavior, a framework of innovation arenas is proposed. It supports the analysis of ongoing E-Government projects as well as the ex post investigation of successful or failed projects. Testing this framework in case studies already demonstrates its usefulness for individual actors making strategic choices about change management. Furthermore, the results indicate that many commonly held assumptions about successful change management have to be reconsidered}, language = {en} } @article{MeineckeZieheKurthsetal.2005, author = {Meinecke, Frank C. and Ziehe, Andreas and Kurths, J{\"u}rgen and M{\"u}ller, Klaus-Robert}, title = {Measuring phase synchronization of superimposed signals}, issn = {0031-9007}, year = {2005}, abstract = {Phase synchronization is an important phenomenon that occurs in a wide variety of complex oscillatory processes. Measuring phase synchronization can therefore help to gain fundamental insight into nature. In this Letter we point out that synchronization analysis techniques can detect spurious synchronization, if they are fed with a superposition of signals such as in electroencephalography or magnetoencephalography data. We show how techniques from blind source separation can help to nevertheless measure the true synchronization and avoid such pitfalls}, language = {en} } @article{ScholzKaplanGuyetal.2005, author = {Scholz, Matthias and Kaplan, F. and Guy, C. L. and Kopka, Joachim and Selbig, Joachim}, title = {Non-linear PCA : a missing data approach}, issn = {1367-4803}, year = {2005}, abstract = {Motivation: Visualizing and analysing the potential non-linear structure of a dataset is becoming an important task in molecular biology. This is even more challenging when the data have missing values. Results: Here, we propose an inverse model that performs non-linear principal component analysis (NLPCA) from incomplete datasets. Missing values are ignored while optimizing the model, but can be estimated afterwards. Results are shown for both artificial and experimental datasets. In contrast to linear methods, non-linear methods were able to give better missing value estimations for non-linear structured data. Application: We applied this technique to a time course of metabolite data from a cold stress experiment on the model plant Arabidopsis thaliana, and could approximate the mapping function from any time point to the metabolite responses. Thus, the inverse NLPCA provides greatly improved information for better understanding the complex response to cold stress}, language = {en} } @phdthesis{Nienhaus2005, author = {Nienhaus, Marc}, title = {Real-Time-Non-Photorealistic rendering techniques for illustrating 3D scenes and their dynamics}, address = {Potsdam}, pages = {x, 102 S. : graph. Darst.}, year = {2005}, language = {en} } @article{Camales2005, author = {Camales, Renaud}, title = {Explicit formulation of the solution of Hamada-Leray-Wagschal's theorem}, issn = {0034-5318}, year = {2005}, abstract = {In this paper, an explicit formula of the solution of Hainada-Leray-Wagschal's theorem is given. For this, only structure's theorem of finite dimensional determination's function and linear algebra technics developped in [1] are used}, language = {en} } @article{HaegerSchubert2005, author = {H{\"a}ger, Sebastian and Schubert, Wolfgang}, title = {Assoziationen in Softwarearchitekturen}, series = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, volume = {2005, 2}, journal = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, publisher = {Univ.}, address = {Potsdam}, issn = {0946-7580}, pages = {33 S.}, year = {2005}, language = {de} } @article{AngerGebserLinkeetal.2005, author = {Anger, Christian and Gebser, Martin and Linke, Thomas and Neumann, Andre and Schaub, Torsten H.}, title = {The nomore++ approach to answer set solving}, year = {2005}, language = {en} } @book{Boerner2005, author = {B{\"o}rner, Ferdinand}, title = {Kompaktierung von Daten aus Scanpfaden mit einem Faltungskompaktor}, series = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, volume = {2005, 3}, journal = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, publisher = {Univ.}, address = {Potsdam}, issn = {0946-7580}, pages = {39 S.}, year = {2005}, language = {de} } @article{GressmannJanhunenMerceretal.2005, author = {Gressmann, Jean and Janhunen, Tomi and Mercer, Robert E. and Schaub, Torsten H. and Thiele, Sven and Tichy, Richard}, title = {Platypus : a platform for distributed answer set solving}, year = {2005}, language = {en} } @article{AngerGebserLinkeetal.2005, author = {Anger, Christian and Gebser, Martin and Linke, Thomas and Neumann, Andre and Schaub, Torsten H.}, title = {The nomore++ approach to answer set solving}, year = {2005}, language = {en} } @article{GrellKonczakSchaub2005, author = {Grell, Susanne and Konczak, Kathrin and Schaub, Torsten H.}, title = {nomore) : a system for computing preferred Answer Sets}, issn = {0302-9743}, year = {2005}, language = {en} } @article{FaberKonczak2005, author = {Faber, Wolfgang and Konczak, Kathrin}, title = {Strong Equivalence for Logic Programs with Preferences}, year = {2005}, language = {en} } @article{AngerKonczakLinkeetal.2005, author = {Anger, Christian and Konczak, Kathrin and Linke, Thomas and Schaub, Torsten H.}, title = {A Glimpse of Answer Set Programming}, issn = {0170-4516}, year = {2005}, language = {en} } @article{KonczakLang2005, author = {Konczak, Kathrin and Lang, Jerome}, title = {Voting procedures with incomplete preferences}, year = {2005}, language = {en} } @article{KonczakVogel2005, author = {Konczak, Kathrin and Vogel, Ralf}, title = {Abduction and preferences in linguistics : Extended abstract}, issn = {0302-9743}, year = {2005}, language = {en} } @article{GoesselMorozovSapozhnikovetal.2005, author = {Goessel, Michael and Morozov, A. V. and Sapozhnikov, V. V. and Sapozhaikov, Vl. V.}, title = {Checking combinational circuits by the method of logic complement}, issn = {0005-1179}, year = {2005}, abstract = {Design of fully self-testing combinational circuits was considered. A theorem defining the conditions for guaranteed logic complement-based design of fully self-testing circuit was proved. Examples were presented}, language = {en} } @article{DelgrandeSchaub2005, author = {Delgrande, James Patrick and Schaub, Torsten H.}, title = {Expressing default logic variants in default logic}, issn = {0955-792X}, year = {2005}, abstract = {Reiter's default logic is one of the best known and most studied of the approaches to nonmonotonic reasoning. Several variants of default logic have subsequently been proposed to give systems with properties differing from the original. In this paper, we examine the relationship between default logic and its major variants. We accomplish this by translating a default theory under a variant interpretation into a second default theory, under the original Reiter semantics, wherein the variant interpretation is respected. That is, in each case we show that, given an extension of a translated theory, one may extract an extension of the original variant default logic theory. We show how constrained, rational, justified, and cumulative default logic can be expressed in Reiter's default logic. As well, we show how Reiter's default logic can be expressed in rational default logic. From this, we suggest that any such variant can be similarly treated. Consequently, we provide a unification of default logics, showing how the original formulation of default logic may express its variants. Moreover, the translations clearly express the relationships between alternative approaches to default logic. The translations themselves are shown to generally have good properties. Thus, in at least a theoretical sense, we show that these variants are in a sense superfluous, in that for any of these variants of default logic, we can exactly mimic the behaviour of a variant in standard default logic. As well, the translations lend insight into means of classifying the expressive power of default logic variants; specifically we suggest that the property of semi-monotonicity represents a division with respect to expressibility, whereas regularity and cumulativity do not}, language = {en} } @article{MeineckeHarmelingMueller2005, author = {Meinecke, Frank C. and Harmeling, Stefan and M{\"u}ller, Klaus-Robert}, title = {Inlier-based ICA with an application to superimposed images}, issn = {0899-9457}, year = {2005}, abstract = {This paper proposes a new independent component analysis (ICA) method which is able to unmix overcomplete mixtures of sparce or structured signals like speech, music or images. Furthermore, the method is designed to be robust against outliers, which is a favorable feature for ICA algorithms since most of them are extremely sensitive to outliers. Our approach is based on a simple outlier index. However, instead of robustifying an existing algorithm by some outlier rejection technique we show how this index can be used directly to solve the ICA problem for super-Gaussian sources. The resulting inlier-based ICA (IBICA) is outlier-robust by construction and can be used for standard ICA as well as for overcomplete ICA (i.e. more source signals than observed signals). (c) 2005 Wiley Periodicals, Inc}, language = {en} } @article{LemmBlankertzCurioetal.2005, author = {Lemm, Steven and Blankertz, Benjamin and Curio, Gabriel and M{\"u}ller, Klaus-Robert}, title = {Spatio-spectral filters for improving the classification of single trial EEG}, issn = {0018-9294}, year = {2005}, abstract = {Data recorded in electroencephalogram (EEG)-based brain-computer interface experiments is generally very noisy, non-stationary, and contaminated with artifacts that can deteriorate discrimination/classification methods. In this paper, we extend the common spatial pattern (CSP) algorithm with the aim to alleviate these adverse effects. In particular, we suggest an extension of CSP to the state space, which utilizes the method of time delay embedding. As we will show, this allows for individually tuned frequency filters at each electrode position and, thus, yields an improved and more robust machine learning procedure. The advantages of the proposed method over the original CSP method are verified in terms of an improved information transfer rate (bits per trial) on a set of EEG-recordings from experiments of imagined limb movements}, language = {en} } @article{WilligMatheusWolisz2005, author = {Willig, Andreas and Matheus, K. and Wolisz, A.}, title = {Wireless technology in industrial networks}, issn = {0018-9219}, year = {2005}, abstract = {With the success of wireless technologies in consumer electronics, standard wireless technologies are envisioned for the deployment in industrial environments as well. Industrial applications involving mobile subsystems or just the desire to save cabling make wireless technologies attractive. Nevertheless, these applications often have stringent requirements on reliability and timing. In wired environments, timing and reliability are well catered for by fieldbus systems (which are a mature technology designed to enable communication between digital controllers and the sensors and actuators interfacing to a physical process). When wireless links are included, reliability and timing requirements are significantly more difficult to meet, due to the adverse properties of the radio channels. In this paper we thus discuss some key issues coming up in wireless fieldbus and wireless industrial communication systems:1)fundamental problems like achieving timely and reliable transmission despite channel errors; 2) the usage of existing wireless technologies for this specific field of applications; and 3) the creation of hybrid systems in which wireless stations are included into existing wired systems}, language = {en} } @phdthesis{RobinsonMallett2005, author = {Robinson-Mallett, Christopher}, title = {Modellbasierte Modulpr{\"u}fung f{\"u}r die Entwicklung technischer, softwareintensiver Systeme mit Real-Time Object-Oriented Modeling}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-6045}, school = {Universit{\"a}t Potsdam}, year = {2005}, abstract = {Mit zunehmender Komplexit{\"a}t technischer Softwaresysteme ist die Nachfrage an produktiveren Methoden und Werkzeugen auch im sicherheitskritischen Umfeld gewachsen. Da insbesondere objektorientierte und modellbasierte Ans{\"a}tze und Methoden ausgezeichnete Eigenschaften zur Entwicklung großer und komplexer Systeme besitzen, ist zu erwarten, dass diese in naher Zukunft selbst bis in sicherheitskritische Bereiche der Softwareentwicklung vordringen. Mit der Unified Modeling Language Real-Time (UML-RT) wird eine Softwareentwicklungsmethode f{\"u}r technische Systeme durch die Object Management Group (OMG) propagiert. F{\"u}r den praktischen Einsatz im technischen und sicherheitskritischen Umfeld muss diese Methode nicht nur bestimmte technische Eigenschaften, beispielsweise temporale Analysierbarkeit, besitzen, sondern auch in einen bestehenden Qualit{\"a}tssicherungsprozess integrierbar sein. Ein wichtiger Aspekt der Integration der UML-RT in ein qualit{\"a}tsorientiertes Prozessmodell, beispielsweise in das V-Modell, ist die Verf{\"u}gbarkeit von ausgereiften Konzepten und Methoden f{\"u}r einen systematischen Modultest. Der Modultest dient als erste Qualitit{\"a}tssicherungsphase nach der Implementierung der Fehlerfindung und dem Qualit{\"a}tsnachweis f{\"u}r jede separat pr{\"u}fbare Softwarekomponente eines Systems. W{\"a}hrend dieser Phase stellt die Durchf{\"u}hrung von systematischen Tests die wichtigste Qualit{\"a}tssicherungsmaßnahme dar. W{\"a}hrend zum jetzigen Zeitpunkt zwar ausgereifte Methoden und Werkzeuge f{\"u}r die modellbasierte Softwareentwicklung zur Verf{\"u}gung stehen, existieren nur wenig {\"u}berzeugende L{\"o}sungen f{\"u}r eine systematische modellbasierte Modulpr{\"u}fung. Die durchg{\"a}ngige Verwendung ausf{\"u}hrbarer Modelle und Codegenerierung stellen wesentliche Konzepte der modellbasierten Softwareentwicklung dar. Sie dienen der konstruktiven Fehlerreduktion durch Automatisierung ansonsten fehlertr{\"a}chtiger, manueller Vorg{\"a}nge. Im Rahmen einer modellbasierten Qualit{\"a}tssicherung sollten diese Konzepte konsequenterweise in die sp{\"a}teren Qualit{\"a}tssicherungsphasen transportiert werden. Daher ist eine wesentliche Forderung an ein Verfahren zur modellbasierten Modulpr{\"u}fung ein m{\"o}glichst hoher Grad an Automatisierung. In aktuellen Entwicklungen hat sich f{\"u}r die Generierung von Testf{\"a}llen auf Basis von Zustandsautomaten die Verwendung von Model Checking als effiziente und an die vielf{\"a}ltigsten Testprobleme anpassbare Methode bew{\"a}hrt. Der Ansatz des Model Checking stammt urspr{\"u}nglich aus dem Entwurf von Kommunikationsprotokollen und wurde bereits erfolgreich auf verschiedene Probleme der Modellierung technischer Software angewendet. Insbesondere in der Gegenwart ausf{\"u}hrbarer, automatenbasierter Modelle erscheint die Verwendung von Model Checking sinnvoll, das die Existenz einer formalen, zustandsbasierten Spezifikation voraussetzt. Ein ausf{\"u}hrbares, zustandsbasiertes Modell erf{\"u}llt diese Anforderungen in der Regel. Aus diesen Gr{\"u}nden ist die Wahl eines Model Checking Ansatzes f{\"u}r die Generierung von Testf{\"a}llen im Rahmen eines modellbasierten Modultestverfahrens eine logische Konsequenz. Obwohl in der aktuellen Spezifikation der UML-RT keine eindeutigen Aussagen {\"u}ber den zur Verhaltensbeschreibung zu verwendenden Formalismus gemacht werden, ist es wahrscheinlich, dass es sich bei der UML-RT um eine zu Real-Time Object-Oriented Modeling (ROOM) kompatible Methode handelt. Alle in dieser Arbeit pr{\"a}sentierten Methoden und Ergebnisse sind somit auf die kommende UML-RT {\"u}bertragbar und von sehr aktueller Bedeutung. Aus den genannten Gr{\"u}nden verfolgt diese Arbeit das Ziel, die analytische Qualit{\"a}tssicherung in der modellbasierten Softwareentwicklung mittels einer modellbasierten Methode f{\"u}r den Modultest zu verbessern. Zu diesem Zweck wird eine neuartige Testmethode pr{\"a}sentiert, die auf automatenbasierten Verhaltensmodellen und CTL Model Checking basiert. Die Testfallgenerierung kann weitgehend automatisch erfolgen, um Fehler durch menschlichen Einfluss auszuschließen. Das entwickelte Modultestverfahren ist in die technischen Konzepte Model Driven Architecture und ROOM, beziehungsweise UML-RT, sowie in die organisatorischen Konzepte eines qualit{\"a}tsorientierten Prozessmodells, beispielsweise das V-Modell, integrierbar.}, subject = {Software}, language = {de} } @phdthesis{Ziehe2005, author = {Ziehe, Andreas}, title = {Blind source separation based on joint diagonalization of matrices with applications in biomedical signal processing}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-5694}, school = {Universit{\"a}t Potsdam}, year = {2005}, abstract = {This thesis is concerned with the solution of the blind source separation problem (BSS). The BSS problem occurs frequently in various scientific and technical applications. In essence, it consists in separating meaningful underlying components out of a mixture of a multitude of superimposed signals. In the recent research literature there are two related approaches to the BSS problem: The first is known as Independent Component Analysis (ICA), where the goal is to transform the data such that the components become as independent as possible. The second is based on the notion of diagonality of certain characteristic matrices derived from the data. Here the goal is to transform the matrices such that they become as diagonal as possible. In this thesis we study the latter method of approximate joint diagonalization (AJD) to achieve a solution of the BSS problem. After an introduction to the general setting, the thesis provides an overview on particular choices for the set of target matrices that can be used for BSS by joint diagonalization. As the main contribution of the thesis, new algorithms for approximate joint diagonalization of several matrices with non-orthogonal transformations are developed. These newly developed algorithms will be tested on synthetic benchmark datasets and compared to other previous diagonalization algorithms. Applications of the BSS methods to biomedical signal processing are discussed and exemplified with real-life data sets of multi-channel biomagnetic recordings.}, subject = {Signaltrennung}, language = {en} } @phdthesis{Floeter2005, author = {Fl{\"o}ter, Andr{\´e}}, title = {Analyzing biological expression data based on decision tree induction}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-6416}, school = {Universit{\"a}t Potsdam}, year = {2005}, abstract = {Modern biological analysis techniques supply scientists with various forms of data. One category of such data are the so called "expression data". These data indicate the quantities of biochemical compounds present in tissue samples. Recently, expression data can be generated at a high speed. This leads in turn to amounts of data no longer analysable by classical statistical techniques. Systems biology is the new field that focuses on the modelling of this information. At present, various methods are used for this purpose. One superordinate class of these meth­ods is machine learning. Methods of this kind had, until recently, predominantly been used for classification and prediction tasks. This neglected a powerful secondary benefit: the ability to induce interpretable models. Obtaining such models from data has become a key issue within Systems biology. Numerous approaches have been proposed and intensively discussed. This thesis focuses on the examination and exploitation of one basic technique: decision trees. The concept of comparing sets of decision trees is developed. This method offers the pos­sibility of identifying significant thresholds in continuous or discrete valued attributes through their corresponding set of decision trees. Finding significant thresholds in attributes is a means of identifying states in living organisms. Knowing about states is an invaluable clue to the un­derstanding of dynamic processes in organisms. Applied to metabolite concentration data, the proposed method was able to identify states which were not found with conventional techniques for threshold extraction. A second approach exploits the structure of sets of decision trees for the discovery of com­binatorial dependencies between attributes. Previous work on this issue has focused either on expensive computational methods or the interpretation of single decision trees ­ a very limited exploitation of the data. This has led to incomplete or unstable results. That is why a new method is developed that uses sets of decision trees to overcome these limitations. Both the introduced methods are available as software tools. They can be applied consecu­tively or separately. That way they make up a package of analytical tools that usefully supplement existing methods. By means of these tools, the newly introduced methods were able to confirm existing knowl­edge and to suggest interesting and new relationships between metabolites.}, subject = {Molekulare Bioinformatik}, language = {en} } @article{Doellner2005, author = {D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Geospatial digital rights management in geovisualization}, issn = {0008-7041}, year = {2005}, abstract = {Geovisualization offers powerful tools, techniques, and strategies to present, explore, analyze, and manage geoinformation. Interactive geovirtual environments such as virtual 3D maps or virtual 3D city models, however, raise the question how to control geodata usage and distribution. We present a concept for embedding digital rights in geovisualizations. It is based on geo-documents, an object-oriented scheme to specify a wide range of geo visualizations. Geo-documents are assembled by building blocks categorized into presentation, structure, interaction, animation, and Digital Rights Management (DRM) classes. DRM objects allow for defining permissions and constraints for all objects contained in geo-documents. In this way, authors of geo visualizations can control how their geo-documents are used, personalized, and redistributed by users. The strengths of the presented concept include the ability to integrate heterogeneous 2D and 3D geodata within a compact design scheme and the ability to cope with privacy, security, and copyright issues. Embedded digital rights in geovisualizations can be applied to improve the usability of geodata user interfaces, to implement publisher-subscriber communication systems for geodata, and to establish business models for geodata trading systems}, language = {en} } @book{KirschNienhausDoellner2005, author = {Kirsch, Florian and Nienhaus, Marc and D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Visualizing design and spatial assembly of interactive CSG}, series = {Technische Berichte des Hasso-Plattner-Instituts f{\"u}r Softwaresystemtechnik an der Universit{\"a}t Potsda}, volume = {7}, journal = {Technische Berichte des Hasso-Plattner-Instituts f{\"u}r Softwaresystemtechnik an der Universit{\"a}t Potsda}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, isbn = {3-937786-56-2}, issn = {1613-5652}, pages = {8 S.}, year = {2005}, language = {en} } @article{NienhausDoellner2005, author = {Nienhaus, Marc and D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Depicting dynamics using principles of visual art and narration's}, issn = {0272-1716}, year = {2005}, language = {en} } @phdthesis{Morozov2005, author = {Morozov, Alexei}, title = {Optimierung von Fehlererkennungsschaltungen auf der Grundlage von komplement{\"a}ren Erg{\"a}nzungen f{\"u}r 1-aus-3 und Berger Codes}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-5360}, school = {Universit{\"a}t Potsdam}, year = {2005}, abstract = {Die Dissertation stellt eine neue Herangehensweise an die L{\"o}sung der Aufgabe der funktionalen Diagnostik digitaler Systeme vor. In dieser Arbeit wird eine neue Methode f{\"u}r die Fehlererkennung vorgeschlagen, basierend auf der Logischen Erg{\"a}nzung und der Verwendung von Berger-Codes und dem 1-aus-3 Code. Die neue Fehlererkennungsmethode der Logischen Erg{\"a}nzung gestattet einen hohen Optimierungsgrad der ben{\"o}tigten Realisationsfl{\"a}che der konstruierten Fehlererkennungsschaltungen. Außerdem ist eins der wichtigen in dieser Dissertation gel{\"o}sten Probleme die Synthese vollst{\"a}ndig selbstpr{\"u}fender Schaltungen.}, subject = {logische Erg{\"a}nzung}, language = {de} } @phdthesis{Kirsch2005, author = {Kirsch, Florian}, title = {Entwurf und Implementierung eines computergraphischen Systems zur Integration komplexer, echtzeitf{\"a}higer 3D-Renderingverfahren}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-6079}, school = {Universit{\"a}t Potsdam}, year = {2005}, abstract = {Thema dieser Arbeit sind echtzeitf{\"a}hige 3D-Renderingverfahren, die 3D-Geometrie mit {\"u}ber der Standarddarstellung hinausgehenden Qualit{\"a}ts- und Gestaltungsmerkmalen rendern k{\"o}nnen. Beispiele sind Verfahren zur Darstellung von Schatten, Reflexionen oder Transparenz. Mit heutigen computergraphischen Software-Basissystemen ist ihre Integration in 3D-Anwendungssysteme sehr aufw{\"a}ndig: Dies liegt einerseits an der technischen, algorithmischen Komplexit{\"a}t der Einzelverfahren, andererseits an Ressourcenkonflikten und Seiteneffekten bei der Kombination mehrerer Verfahren. Szenengraphsysteme, intendiert als computergraphische Softwareschicht zur Abstraktion von der Graphikhardware, stellen derzeit keine Mechanismen zur Nutzung dieser Renderingverfahren zur Verf{\"u}gung. Ziel dieser Arbeit ist es, eine Software-Architektur f{\"u}r ein Szenengraphsystem zu konzipieren und umzusetzen, die echtzeitf{\"a}hige 3D-Renderingverfahren als Komponenten modelliert und es damit erlaubt, diese Verfahren innerhalb des Szenengraphsystems f{\"u}r die Anwendungsentwicklung effektiv zu nutzen. Ein Entwickler, der ein solches Szenengraphsystem nutzt, steuert diese Komponenten durch Elemente in der Szenenbeschreibung an, die die sichtbare Wirkung eines Renderingverfahrens auf die Geometrie in der Szene angeben, aber keine Hinweise auf die algorithmische Implementierung des Verfahrens enthalten. Damit werden Renderingverfahren in 3D-Anwendungssystemen nutzbar, ohne dass ein Entwickler detaillierte Kenntnisse {\"u}ber sie ben{\"o}tigt, so dass der Aufwand f{\"u}r ihre Entwicklung drastisch reduziert wird. Ein besonderer Augenmerk der Arbeit liegt darauf, auf diese Weise auch verschiedene Renderingverfahren in einer Szene kombiniert einsetzen zu k{\"o}nnen. Hierzu ist eine Unterteilung der Renderingverfahren in mehrere Kategorien erforderlich, die mit Hilfe unterschiedlicher Ans{\"a}tze ausgewertet werden. Dies erlaubt die Abstimmung verschiedener Komponenten f{\"u}r Renderingverfahren und ihrer verwendeten Ressourcen. Die Zusammenarbeit mehrerer Renderingverfahren hat dort ihre Grenzen, wo die Kombination von Renderingverfahren graphisch nicht sinnvoll ist oder fundamentale technische Beschr{\"a}nkungen der Verfahren eine gleichzeitige Verwendung unm{\"o}glich machen. Die in dieser Arbeit vorgestellte Software-Architektur kann diese Grenzen nicht verschieben, aber sie erm{\"o}glicht den gleichzeitigen Einsatz vieler Verfahren, bei denen eine Kombination aufgrund der hohen Komplexit{\"a}t der Implementierung bislang nicht erreicht wurde. Das Verm{\"o}gen zur Zusammenarbeit ist dabei allerdings von der Art eines Einzelverfahrens abh{\"a}ngig: Verfahren zur Darstellung transparenter Geometrie beispielsweise erfordern bei der Kombination mit anderen Verfahren in der Regel vollst{\"a}ndig neuentwickelte Renderingverfahren; entsprechende Komponenten f{\"u}r das Szenengraphsystem k{\"o}nnen daher nur eingeschr{\"a}nkt mit Komponenten f{\"u}r andere Renderingverfahren verwendet werden. Das in dieser Arbeit entwickelte System integriert und kombiniert Verfahren zur Darstellung von Bumpmapping, verschiedene Schatten- und Reflexionsverfahren sowie bildbasiertes CSG-Rendering. Damit stehen wesentliche Renderingverfahren in einem Szenengraphsystem erstmalig komponentenbasiert und auf einem hohen Abstraktionsniveau zur Verf{\"u}gung. Das System ist trotz des zus{\"a}tzlichen Verwaltungsaufwandes in der Lage, die Renderingverfahren einzeln und in Kombination grunds{\"a}tzlich in Echtzeit auszuf{\"u}hren.}, subject = {Dreidimensionale Computergraphik}, language = {de} } @article{SchefflerSchnor2005, author = {Scheffler, Thomas and Schnor, Bettina}, title = {Privacy Requirements for Embedded Sensor Devices}, isbn = {978-3-800729-09-8}, year = {2005}, abstract = {This paper analyses data privacy issues as they arise from different deployment scenarios for networks that use embedded sensor devices. Maintaining data privacy in pervasive environments requires the management and implementation of privacy protection measures close to the data source. We propose a set of atomic privacy parameters that is generic enough to form specific privacy classes and might be applied directly at the embedded sensor device.}, language = {en} } @article{SchneidenbachSchnor2005, author = {Schneidenbach, Lars and Schnor, Bettina}, title = {Migration of MPI Applications to IPv6 Networks}, isbn = {0-88986-468-3}, year = {2005}, language = {en} } @article{FriedrichSchneidenbachSchnor2005, author = {Friedrich, Sven and Schneidenbach, Lars and Schnor, Bettina}, title = {SLIBNet : Server Load Balancing for InfiniBand Networks}, year = {2005}, abstract = {Today, InfiniBand is an evolving high speed interconnect technology to build high performance computing clusters, that achieve top 10 rankings in the current top 500 of the worldwide fastest supercomputers. Network interfaces (called host channel adapters) provide transport layer services over connections and datagrams in reliable or unreliable manner. Additionally, InfiniBand supports remote direct memory access (RDMA) primitives that allow for one- sided communication. Using server load balancing together with a high performance cluster makes it possible to build a fast, scalable, and reliable service infrastructure. We have designed and implemented a scalable load balancer for InfiniBand clusters called SLIBNet. Our investigations show that the InfiniBand architecture offers features which perfectly support load balancing. We want to thank the Megware Computer GmbH for providing us an InfiniBand switch to realize a server load balancing testbed.}, language = {en} } @phdthesis{Ghasemzadeh2005, author = {Ghasemzadeh, Mohammad}, title = {A new algorithm for the quantified satisfiability problem, based on zero-suppressed binary decision diagrams and memoization}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-6378}, school = {Universit{\"a}t Potsdam}, year = {2005}, abstract = {Quantified Boolean formulas (QBFs) play an important role in theoretical computer science. QBF extends propositional logic in such a way that many advanced forms of reasoning can be easily formulated and evaluated. In this dissertation we present our ZQSAT, which is an algorithm for evaluating quantified Boolean formulas. ZQSAT is based on ZBDD: Zero-Suppressed Binary Decision Diagram , which is a variant of BDD, and an adopted version of the DPLL algorithm. It has been implemented in C using the CUDD: Colorado University Decision Diagram package. The capability of ZBDDs in storing sets of subsets efficiently enabled us to store the clauses of a QBF very compactly and let us to embed the notion of memoization to the DPLL algorithm. These points led us to implement the search algorithm in such a way that we could store and reuse the results of all previously solved subformulas with a little overheads. ZQSAT can solve some sets of standard QBF benchmark problems (known to be hard for DPLL based algorithms) faster than the best existing solvers. In addition to prenex-CNF, ZQSAT accepts prenex-NNF formulas. We show and prove how this capability can be exponentially beneficial.}, subject = {Bin{\"a}res Entscheidungsdiagramm}, language = {en} }