@article{HornKoutzevlovRaetschetal.1996, author = {Horn, Erika and Koutzevlov, Atanas and R{\"a}tsch, Gunnar and Schubert, Wolfgang and Tschapek, Alexej}, title = {Software system specification}, series = {HGG-Dokument}, volume = {UPHGG.010.4}, journal = {HGG-Dokument}, publisher = {Univ.}, address = {Potsdam}, year = {1996}, language = {en} } @book{KoutzevlovSchubertTschapek1996, author = {Koutzevlov, Atanas and Schubert, Wolfgang and Tschapek, Alexej}, title = {Software test plan \& test description}, series = {HGG-Dokument}, volume = {UPHGG.016.1}, journal = {HGG-Dokument}, publisher = {Univ.}, address = {Potsdam}, year = {1996}, language = {en} } @article{HornReinkeKupries1999, author = {Horn, Erika and Reinke, Thomas and Kupries, Mario}, title = {Software-Agentsysteme : Konzepte, Anwendungsgebiete, Nutzen}, isbn = {3-929642-28-x}, year = {1999}, language = {de} } @article{ChenLangeAndjelkovicetal.2022, author = {Chen, Junchao and Lange, Thomas and Andjelkovic, Marko and Simevski, Aleksandar and Lu, Li and Krstic, Milos}, title = {Solar particle event and single event upset prediction from SRAM-based monitor and supervised machine learning}, series = {IEEE transactions on emerging topics in computing / IEEE Computer Society, Institute of Electrical and Electronics Engineers}, volume = {10}, journal = {IEEE transactions on emerging topics in computing / IEEE Computer Society, Institute of Electrical and Electronics Engineers}, number = {2}, publisher = {Institute of Electrical and Electronics Engineers}, address = {[New York, NY]}, issn = {2168-6750}, doi = {10.1109/TETC.2022.3147376}, pages = {564 -- 580}, year = {2022}, abstract = {The intensity of cosmic radiation may differ over five orders of magnitude within a few hours or days during the Solar Particle Events (SPEs), thus increasing for several orders of magnitude the probability of Single Event Upsets (SEUs) in space-borne electronic systems. Therefore, it is vital to enable the early detection of the SEU rate changes in order to ensure timely activation of dynamic radiation hardening measures. In this paper, an embedded approach for the prediction of SPEs and SRAM SEU rate is presented. The proposed solution combines the real-time SRAM-based SEU monitor, the offline-trained machine learning model and online learning algorithm for the prediction. With respect to the state-of-the-art, our solution brings the following benefits: (1) Use of existing on-chip data storage SRAM as a particle detector, thus minimizing the hardware and power overhead, (2) Prediction of SRAM SEU rate one hour in advance, with the fine-grained hourly tracking of SEU variations during SPEs as well as under normal conditions, (3) Online optimization of the prediction model for enhancing the prediction accuracy during run-time, (4) Negligible cost of hardware accelerator design for the implementation of selected machine learning model and online learning algorithm. The proposed design is intended for a highly dependable and self-adaptive multiprocessing system employed in space applications, allowing to trigger the radiation mitigation mechanisms before the onset of high radiation levels.}, language = {en} } @article{SchulmeisterRoseVoigt1996, author = {Schulmeister, Thomas and Rose, J{\"u}rgen and Voigt, Horst}, title = {Some extensions of Schmidt's partition method for sequence comparison based on binary character amino acid properties}, year = {1996}, language = {en} } @article{LemmBlankertzCurioetal.2005, author = {Lemm, Steven and Blankertz, Benjamin and Curio, Gabriel and M{\"u}ller, Klaus-Robert}, title = {Spatio-spectral filters for improving the classification of single trial EEG}, issn = {0018-9294}, year = {2005}, abstract = {Data recorded in electroencephalogram (EEG)-based brain-computer interface experiments is generally very noisy, non-stationary, and contaminated with artifacts that can deteriorate discrimination/classification methods. In this paper, we extend the common spatial pattern (CSP) algorithm with the aim to alleviate these adverse effects. In particular, we suggest an extension of CSP to the state space, which utilizes the method of time delay embedding. As we will show, this allows for individually tuned frequency filters at each electrode position and, thus, yields an improved and more robust machine learning procedure. The advantages of the proposed method over the original CSP method are verified in terms of an improved information transfer rate (bits per trial) on a set of EEG-recordings from experiments of imagined limb movements}, language = {en} } @unpublished{LuckeSteinmetz2014, author = {Lucke, Ulrike and Steinmetz, Ralf}, title = {Special issue on "Pervasive Education"}, series = {Pervasive and mobile computing}, volume = {14}, journal = {Pervasive and mobile computing}, publisher = {Elsevier}, address = {Amsterdam}, issn = {1574-1192}, doi = {10.1016/j.pmcj.2014.08.001}, pages = {1 -- 2}, year = {2014}, language = {en} } @misc{SchaubWoltran2018, author = {Schaub, Torsten H. and Woltran, Stefan}, title = {Special issue on answer set programming}, series = {K{\"u}nstliche Intelligenz}, volume = {32}, journal = {K{\"u}nstliche Intelligenz}, number = {2-3}, publisher = {Springer}, address = {Heidelberg}, issn = {0933-1875}, doi = {10.1007/s13218-018-0554-8}, pages = {101 -- 103}, year = {2018}, language = {en} } @article{Bobda2009, author = {Bobda, Christophe}, title = {Special issue on ReCoSoC 2007 : editorial}, issn = {0141-9331}, doi = {10.1016/j.micpro.2009.01.001}, year = {2009}, language = {en} } @article{LiskeRebensburgSchnor2007, author = {Liske, Stefan and Rebensburg, Klaus and Schnor, Bettina}, title = {SPIT-Erkennung, -Bekanntgabe und -Abwehr in SIP-Netzwerken}, isbn = {978-3-540-69961-3}, year = {2007}, abstract = {SPAM ist in den letzten Jahren zur großten Bedrohung der E-Mail-Kommunikation herangewachsen - jedoch nicht nur auf diesen Kommunikationsweg beschrankt. Mit steigender Anzahl von VoIP-Anschl{\"u}ssen werden auch hier die teilnehmenden Benutzer mit SPAM-Anrufen (SPIT) konfrontiert werden. Neben derzeit diskutierten juristischen Maßnahmen m{\"u}ssen auch technische Abwehrmaßnahmen geschaffen werden, welche SPAM erkennen und vermeiden k{\"o}nnen. Dieser Beitrag stellt zwei Erweiterungen f{\"u}r das VoIP-Protokoll SIP vor, welche es erstens den Providern erm{\"o}glichen, SPIT-Einsch{\"a}tzungen {\"u}ber den Anrufer zum angerufenen Benutzer zu {\"u}bermitteln und zweitens den Angerufenen die M{\"o}glichkeit geben, mit einer Kostenanforderung auf potentielle SPIT-Anrufe zu reagieren.}, language = {de} } @article{Hibbe2014, author = {Hibbe, Marcel}, title = {Spotlocator - Guess Where the Photo Was Taken!}, series = {Process Design for Natural Scientists: an agile model-driven approach}, journal = {Process Design for Natural Scientists: an agile model-driven approach}, number = {500}, editor = {Lambrecht, Anna-Lena and Margaria, Tiziana}, publisher = {Springer Verlag}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {149 -- 160}, year = {2014}, abstract = {Spotlocator is a game wherein people have to guess the spots of where photos were taken. The photos of a defined area for each game are from panoramio.com. They are published at http://spotlocator. drupalgardens.com with an ID. Everyone can guess the photo spots by sending a special tweet via Twitter that contains the hashtag \#spotlocator, the guessed coordinates and the ID of the photo. An evaluation is published for all tweets. The players are informed about the distance to the real photo spots and the positions are shown on a map.}, language = {en} } @article{KyprianidisCollomosseWangetal.2013, author = {Kyprianidis, Jan Eric and Collomosse, John and Wang, Tinghuai and Isenberg, Tobias}, title = {State of the "Art" a taxonomy of artistic stylization techniques for images and video}, series = {IEEE transactions on visualization and computer graphics}, volume = {19}, journal = {IEEE transactions on visualization and computer graphics}, number = {5}, publisher = {Inst. of Electr. and Electronics Engineers}, address = {Los Alamitos}, issn = {1077-2626}, doi = {10.1109/TVCG.2012.160}, pages = {866 -- 885}, year = {2013}, abstract = {This paper surveys the field of nonphotorealistic rendering (NPR), focusing on techniques for transforming 2D input (images and video) into artistically stylized renderings. We first present a taxonomy of the 2D NPR algorithms developed over the past two decades, structured according to the design characteristics and behavior of each technique. We then describe a chronology of development from the semiautomatic paint systems of the early nineties, through to the automated painterly rendering systems of the late nineties driven by image gradient analysis. Two complementary trends in the NPR literature are then addressed, with reference to our taxonomy. First, the fusion of higher level computer vision and NPR, illustrating the trends toward scene analysis to drive artistic abstraction and diversity of style. Second, the evolution of local processing approaches toward edge-aware filtering for real-time stylization of images and video. The survey then concludes with a discussion of open challenges for 2D NPR identified in recent NPR symposia, including topics such as user and aesthetic evaluation.}, language = {en} } @article{BruecknerKanzowScheffer2012, author = {Br{\"u}ckner, Michael and Kanzow, Christian and Scheffer, Tobias}, title = {Static prediction games for adversarial learning problems}, series = {Journal of machine learning research}, volume = {13}, journal = {Journal of machine learning research}, publisher = {Microtome Publishing}, address = {Cambridge, Mass.}, issn = {1532-4435}, pages = {2617 -- 2654}, year = {2012}, abstract = {The standard assumption of identically distributed training and test data is violated when the test data are generated in response to the presence of a predictive model. This becomes apparent, for example, in the context of email spam filtering. Here, email service providers employ spam filters, and spam senders engineer campaign templates to achieve a high rate of successful deliveries despite the filters. We model the interaction between the learner and the data generator as a static game in which the cost functions of the learner and the data generator are not necessarily antagonistic. We identify conditions under which this prediction game has a unique Nash equilibrium and derive algorithms that find the equilibrial prediction model. We derive two instances, the Nash logistic regression and the Nash support vector machine, and empirically explore their properties in a case study on email spam filtering.}, language = {en} } @phdthesis{Hentschel2013, author = {Hentschel, Uwe}, title = {Steuerung der Daten{\"u}bertragung in {\"o}ffentlichen zellularen Funknetzen im Kontext telemedizinischer Anwendung}, address = {Potsdam}, pages = {224 S.}, year = {2013}, language = {de} } @article{ThonLandwehrDeRaedt2011, author = {Thon, Ingo and Landwehr, Niels and De Raedt, Luc}, title = {Stochastic relational processes efficient inference and applications}, series = {Machine learning}, volume = {82}, journal = {Machine learning}, number = {2}, publisher = {Springer}, address = {Dordrecht}, issn = {0885-6125}, doi = {10.1007/s10994-010-5213-8}, pages = {239 -- 272}, year = {2011}, abstract = {One of the goals of artificial intelligence is to develop agents that learn and act in complex environments. Realistic environments typically feature a variable number of objects, relations amongst them, and non-deterministic transition behavior. While standard probabilistic sequence models provide efficient inference and learning techniques for sequential data, they typically cannot fully capture the relational complexity. On the other hand, statistical relational learning techniques are often too inefficient to cope with complex sequential data. In this paper, we introduce a simple model that occupies an intermediate position in this expressiveness/efficiency trade-off. It is based on CP-logic (Causal Probabilistic Logic), an expressive probabilistic logic for modeling causality. However, by specializing CP-logic to represent a probability distribution over sequences of relational state descriptions and employing a Markov assumption, inference and learning become more tractable and effective. Specifically, we show how to solve part of the inference and learning problems directly at the first-order level, while transforming the remaining part into the problem of computing all satisfying assignments for a Boolean formula in a binary decision diagram. We experimentally validate that the resulting technique is able to handle probabilistic relational domains with a substantial number of objects and relations.}, language = {en} } @article{LuckeHaferHartmann2020, author = {Lucke, Ulrike and Hafer, J{\"o}rg and Hartmann, Niklas}, title = {Strategieentwicklung in der Hochschule als partizipativer Prozess}, series = {Potsdamer Beitr{\"a}ge zur Hochschulforschung}, journal = {Potsdamer Beitr{\"a}ge zur Hochschulforschung}, number = {6}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, isbn = {978-3-86956-498-2}, issn = {2192-1075}, doi = {10.25932/publishup-49276}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-492764}, pages = {99 -- 117}, year = {2020}, abstract = {Die Setzung strategischer Ziele sowie die Zuordnung und Umsetzung dazugeh{\"o}render Maßnahmen sind ein wesentliches Element, um die Innovationsf{\"a}higkeit von Organisationen zu erhalten. In den vergangenen Jahren ist auch an Hochschulen die Strategiebildung deutlich vorangetrieben worden. Dies betrifft verschiedene Handlungsfelder, und es werden verschiedene Ans{\"a}tze verfolgt. Der vorliegende Beitrag greift am Beispiel der Universit{\"a}t Potsdam drei in den vergangenen Jahren adressierte Strategiebereiche heraus: IT, E-Learning und Forschungsdaten. Die damit verbundenen Prozesse waren in unterschiedlichem Maß von Partizipation gepr{\"a}gt. Die gesammelten Erfahrungen werden reflektiert, und es werden Empfehlungen f{\"u}r Strategieentwicklungsprozesse abgeleitet.}, language = {de} } @article{FaberKonczak2005, author = {Faber, Wolfgang and Konczak, Kathrin}, title = {Strong Equivalence for Logic Programs with Preferences}, year = {2005}, language = {en} } @book{StoffelKunz1997, author = {Stoffel, Dominik and Kunz, Wolfgang}, title = {Structural FSM traversal : theory and a practical algorithm}, series = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, volume = {1997, 05}, journal = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, publisher = {Univ. Potsdam}, address = {Potsdam}, issn = {0946-7580}, pages = {34 S. : Ill.}, year = {1997}, language = {en} } @phdthesis{Kyprianidis2013, author = {Kyprianidis, Jan Eric}, title = {Structure adaptive stylization of images and video}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-64104}, school = {Universit{\"a}t Potsdam}, year = {2013}, abstract = {In the early days of computer graphics, research was mainly driven by the goal to create realistic synthetic imagery. By contrast, non-photorealistic computer graphics, established as its own branch of computer graphics in the early 1990s, is mainly motivated by concepts and principles found in traditional art forms, such as painting, illustration, and graphic design, and it investigates concepts and techniques that abstract from reality using expressive, stylized, or illustrative rendering techniques. This thesis focuses on the artistic stylization of two-dimensional content and presents several novel automatic techniques for the creation of simplified stylistic illustrations from color images, video, and 3D renderings. Primary innovation of these novel techniques is that they utilize the smooth structure tensor as a simple and efficient way to obtain information about the local structure of an image. More specifically, this thesis contributes to knowledge in this field in the following ways. First, a comprehensive review of the structure tensor is provided. In particular, different methods for integrating the minor eigenvector field of the smoothed structure tensor are developed, and the superiority of the smoothed structure tensor over the popular edge tangent flow is demonstrated. Second, separable implementations of the popular bilateral and difference of Gaussians filters that adapt to the local structure are presented. These filters avoid artifacts while being computationally highly efficient. Taken together, both provide an effective way to create a cartoon-style effect. Third, a generalization of the Kuwahara filter is presented that avoids artifacts by adapting the shape, scale, and orientation of the filter to the local structure. This causes directional image features to be better preserved and emphasized, resulting in overall sharper edges and a more feature-abiding painterly effect. In addition to the single-scale variant, a multi-scale variant is presented, which is capable of performing a highly aggressive abstraction. Fourth, a technique that builds upon the idea of combining flow-guided smoothing with shock filtering is presented, allowing for an aggressive exaggeration and an emphasis of directional image features. All presented techniques are suitable for temporally coherent per-frame filtering of video or dynamic 3D renderings, without requiring expensive extra processing, such as optical flow. Moreover, they can be efficiently implemented to process content in real-time on a GPU.}, language = {en} } @phdthesis{Polyvyanyy2012, author = {Polyvyanyy, Artem}, title = {Structuring process models}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-59024}, school = {Universit{\"a}t Potsdam}, year = {2012}, abstract = {One can fairly adopt the ideas of Donald E. Knuth to conclude that process modeling is both a science and an art. Process modeling does have an aesthetic sense. Similar to composing an opera or writing a novel, process modeling is carried out by humans who undergo creative practices when engineering a process model. Therefore, the very same process can be modeled in a myriad number of ways. Once modeled, processes can be analyzed by employing scientific methods. Usually, process models are formalized as directed graphs, with nodes representing tasks and decisions, and directed arcs describing temporal constraints between the nodes. Common process definition languages, such as Business Process Model and Notation (BPMN) and Event-driven Process Chain (EPC) allow process analysts to define models with arbitrary complex topologies. The absence of structural constraints supports creativity and productivity, as there is no need to force ideas into a limited amount of available structural patterns. Nevertheless, it is often preferable that models follow certain structural rules. A well-known structural property of process models is (well-)structuredness. A process model is (well-)structured if and only if every node with multiple outgoing arcs (a split) has a corresponding node with multiple incoming arcs (a join), and vice versa, such that the set of nodes between the split and the join induces a single-entry-single-exit (SESE) region; otherwise the process model is unstructured. The motivations for well-structured process models are manifold: (i) Well-structured process models are easier to layout for visual representation as their formalizations are planar graphs. (ii) Well-structured process models are easier to comprehend by humans. (iii) Well-structured process models tend to have fewer errors than unstructured ones and it is less probable to introduce new errors when modifying a well-structured process model. (iv) Well-structured process models are better suited for analysis with many existing formal techniques applicable only for well-structured process models. (v) Well-structured process models are better suited for efficient execution and optimization, e.g., when discovering independent regions of a process model that can be executed concurrently. Consequently, there are process modeling languages that encourage well-structured modeling, e.g., Business Process Execution Language (BPEL) and ADEPT. However, the well-structured process modeling implies some limitations: (i) There exist processes that cannot be formalized as well-structured process models. (ii) There exist processes that when formalized as well-structured process models require a considerable duplication of modeling constructs. Rather than expecting well-structured modeling from start, we advocate for the absence of structural constraints when modeling. Afterwards, automated methods can suggest, upon request and whenever possible, alternative formalizations that are "better" structured, preferably well-structured. In this thesis, we study the problem of automatically transforming process models into equivalent well-structured models. The developed transformations are performed under a strong notion of behavioral equivalence which preserves concurrency. The findings are implemented in a tool, which is publicly available.}, language = {en} } @phdthesis{Grunske2004, author = {Grunske, Lars}, title = {Strukturorientierte Optimierung der Qualit{\"a}tseigenschaften von softwareintensiven technischen Systemen im Architekturentwurf}, pages = {129 S.}, year = {2004}, language = {de} } @article{EllisAbreuEllis2015, author = {Ellis, Jason Brent and Abreu-Ellis, Carla Reis}, title = {Student Perspectives of Social Networking use in Higher Education}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82632}, pages = {117 -- 131}, year = {2015}, abstract = {Social networks are currently at the forefront of tools that lend to Personal Learning Environments (PLEs). This study aimed to observe how students perceived PLEs, what they believed were the integral components of social presence when using Facebook as part of a PLE, and to describe student's preferences for types of interactions when using Facebook as part of their PLE. This study used mixed methods to analyze the perceptions of graduate and undergraduate students on the use of social networks, more specifically Facebook as a learning tool. Fifty surveys were returned representing a 65 \% response rate. Survey questions included both closed and open-ended questions. Findings suggested that even though students rated themselves relatively well in having requisite technology skills, and 94 \% of students used Facebook primarily for social use, they were hesitant to migrate these skills to academic use because of concerns of privacy, believing that other platforms could fulfil the same purpose, and by not seeing the validity to use Facebook in establishing social presence. What lies at odds with these beliefs is that when asked to identify strategies in Facebook that enabled social presence to occur in academic work, the majority of students identified strategies in five categories that lead to social presence establishment on Facebook during their coursework.}, language = {en} } @article{KlingerPolutinaBibel2013, author = {Klinger, Melanie and Polutina, Olena and Bibel, Ariane}, title = {Studentische eLearning-Beratung}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {5}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-65003}, pages = {131 -- 136}, year = {2013}, abstract = {Der vorliegende Beitrag besch{\"a}ftigt sich mit der Frage, wie der eLearning-Support in großen Institutionen effizient gestaltet werden kann. Vorgestellt wird ein experimentelles Beratungsprojekt, das Lehrende bei der Gestaltung von eLearning-Maßnahmen mithilfe der Lernplattform ILIAS1 unterst{\"u}tzt. Neben der Zielsetzung des Projekts werden dessen Aufbau und erste Praxiserfahrungen er{\"o}rtert. Außerdem werden Potenziale des Beratungsformats, die insbesondere mit der individuellen Vor-Ort-Beratung der Lehrenden durch hochschuldidaktisch geschulte Studierende einhergehen, erl{\"a}utert. Abschließend werden Grenzen und Weiterentwicklungsperspektiven des Projekts dargestellt. Am Beispiel der ILIAS-Beratung soll gezeigt werden, dass es sich einer nachhaltigen Organisationsentwicklung als zutr{\"a}glich erweist, Kooperationen erschiedenartiger Organisationseinheiten zu f{\"o}rdern und die entstehenden Synergieeffekte zu nutzen.}, language = {de} } @article{LeonhardtKwiecienSchmetzetal.2015, author = {Leonhardt, Thiemo and Kwiecien, Alexandra and Schmetz, Arno and Bellgardt, Martin and Naumann, Uwe}, title = {Studienabbruchsquote dauerhaft senken}, series = {HDI 2014 : Gestalten von {\"U}berg{\"a}ngen}, volume = {2015}, journal = {HDI 2014 : Gestalten von {\"U}berg{\"a}ngen}, number = {9}, editor = {Schwill, Andreas and Schubert, Sigrid}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-80274}, pages = {61 -- 76}, year = {2015}, abstract = {Es wird ein umfassendes Mentoring Konzept im Studiengang Informatik an der RWTH Aachen vorgestellt, das den {\"U}bergang von der Schule zur Universit{\"a}t unterst{\"u}tzt und gleichzeitig beim Auftreten von Schwierigkeiten im Verlauf des Studiums effiziente und kompetente Beratung bietet. Das Programm erreicht durchg{\"a}ngig hohe Akzeptanzwerte bei den Studierenden trotz verpflichtender Teilnahme im ersten Semester. Die Wirksamkeit des Programms ist durch die zahlreichen einflussgebenden Variablen zwar rein quantitativ kaum messbar, die M{\"o}glichkeit auf organisatorische und fachliche Probleme eines Jahrgangs reagieren zu k{\"o}nnen sowie einen Einblick auf die Gr{\"u}nde f{\"u}r einen Studienabbruch zu bekommen, best{\"a}tigt aber die Notwendigkeit der Maßnahme.}, language = {de} } @phdthesis{Fudickar2014, author = {Fudickar, Sebastian}, title = {Sub Ghz transceiver for indoor localisation of smartphones}, school = {Universit{\"a}t Potsdam}, pages = {IV, 167}, year = {2014}, language = {en} } @book{TsudaSugiyamaMueller2000, author = {Tsuda, Koji and Sugiyama, Masashi and M{\"u}ller, Klaus-Robert}, title = {Subspace information criterion for non-quadratice regularizers : model selection for sparse regressors}, series = {GMD-Report}, volume = {120}, journal = {GMD-Report}, publisher = {GMD-Forschungszentrum Informationstechnik}, address = {Sankt Augustin}, pages = {36 S.}, year = {2000}, language = {en} } @article{Linke2003, author = {Linke, Thomas}, title = {Suitable graphs for answer set programming}, issn = {1613-0073}, year = {2003}, language = {en} } @article{HempelAdolphsLandwehretal.2020, author = {Hempel, Sabrina and Adolphs, Julian and Landwehr, Niels and Willink, Dilya and Janke, David and Amon, Thomas}, title = {Supervised machine learning to assess methane emissions of a dairy building with natural ventilation}, series = {Applied Sciences}, volume = {10}, journal = {Applied Sciences}, number = {19}, publisher = {MDPI}, address = {Basel}, issn = {2076-3417}, doi = {10.3390/app10196938}, pages = {21}, year = {2020}, abstract = {A reliable quantification of greenhouse gas emissions is a basis for the development of adequate mitigation measures. Protocols for emission measurements and data analysis approaches to extrapolate to accurate annual emission values are a substantial prerequisite in this context. We systematically analyzed the benefit of supervised machine learning methods to project methane emissions from a naturally ventilated cattle building with a concrete solid floor and manure scraper located in Northern Germany. We took into account approximately 40 weeks of hourly emission measurements and compared model predictions using eight regression approaches, 27 different sampling scenarios and four measures of model accuracy. Data normalization was applied based on median and quartile range. A correlation analysis was performed to evaluate the influence of individual features. This indicated only a very weak linear relation between the methane emission and features that are typically used to predict methane emission values of naturally ventilated barns. It further highlighted the added value of including day-time and squared ambient temperature as features. The error of the predicted emission values was in general below 10\%. The results from Gaussian processes, ordinary multilinear regression and neural networks were least robust. More robust results were obtained with multilinear regression with regularization, support vector machines and particularly the ensemble methods gradient boosting and random forest. The latter had the added value to be rather insensitive against the normalization procedure. In the case of multilinear regression, also the removal of not significantly linearly related variables (i.e., keeping only the day-time component) led to robust modeling results. We concluded that measurement protocols with 7 days and six measurement periods can be considered sufficient to model methane emissions from the dairy barn with solid floor with manure scraper, particularly when periods are distributed over the year with a preference for transition periods. Features should be normalized according to median and quartile range and must be carefully selected depending on the modeling approach.}, language = {en} } @article{LewinMcNicol2015, author = {Lewin, Cathy and McNicol, Sarah}, title = {Supporting the Development of 21st Century Skills through ICT}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82672}, pages = {181 -- 198}, year = {2015}, abstract = {The growing impact of globalisation and the development of a 'knowledge society' have led many to argue that 21st century skills are essential for life in twenty-first century society and that ICT is central to their development. This paper describes how 21st century skills, in particular digital literacy, critical thinking, creativity, communication and collaboration skills, have been conceptualised and embedded in the resources developed for teachers in iTEC, a four-year, European project. The effectiveness of this approach is considered in light of the data collected through the evaluation of the pilots, which considers both the potential benefits of using technology to support the development of 21st century skills, but also the challenges of doing so. Finally, the paper discusses the learning support systems required in order to transform pedagogies and embed 21st century skills. It is argued that support is required in standards and assessment; curriculum and instruction; professional development; and learning environments.}, language = {en} } @article{PousttchiGleiss2019, author = {Pousttchi, Key and Gleiß, Alexander}, title = {Surrounded by middlemen - how multi-sided platforms change the insurance industry}, series = {Electron Markets}, volume = {29}, journal = {Electron Markets}, number = {4}, publisher = {Springer}, address = {Heidelberg}, issn = {1019-6781}, doi = {10.1007/s12525-019-00363-w}, pages = {609 -- 629}, year = {2019}, abstract = {Multi-sided platforms (MSP) strongly affect markets and play a crucial part within the digital and networked economy. Although empirical evidence indicates their occurrence in many industries, research has not investigated the game-changing impact of MSP on traditional markets to a sufficient extent. More specifically, we have little knowledge of how MSP affect value creation and customer interaction in entire markets, exploiting the potential of digital technologies to offer new value propositions. Our paper addresses this research gap and provides an initial systematic approach to analyze the impact of MSP on the insurance industry. For this purpose, we analyze the state of the art in research and practice in order to develop a reference model of the value network for the insurance industry. On this basis, we conduct a case-study analysis to discover and analyze roles which are occupied or even newly created by MSP. As a final step, we categorize MSP with regard to their relation to traditional insurance companies, resulting in a classification scheme with four MSP standard types: Competition, Coordination, Cooperation, Collaboration.}, language = {en} } @book{KuropkaMeyer2005, author = {Kuropka, Dominik and Meyer, Harald}, title = {Survey on service composition}, series = {Technische Berichte des Hasso-Plattner-Instituts f{\"u}r Softwaresystemtechnik an der Universit{\"a}t Potsda}, volume = {10}, journal = {Technische Berichte des Hasso-Plattner-Instituts f{\"u}r Softwaresystemtechnik an der Universit{\"a}t Potsda}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, isbn = {3-937786-78-3}, issn = {1613-5652}, pages = {35 S.}, year = {2005}, language = {en} } @book{RaetschSchoelkopfMikaetal.2000, author = {R{\"a}tsch, Gunnar and Sch{\"o}lkopf, B. and Mika, Sebastian and M{\"u}ller, Klaus-Robert}, title = {SVM and boosting : one class}, series = {GMD-Report}, volume = {119}, journal = {GMD-Report}, publisher = {GMD-Forschungszentrum Informationstechnik}, address = {Sankt Augustin}, pages = {36 S.}, year = {2000}, language = {en} } @book{OPUS4-18154, title = {Symbolic and quantitative approaches to reasoning with uncertainty : 6th European conference, ECSQARU 2001, Toulouse, France, September 19 - 21, 2001 ; proceedings}, series = {Lecture notes in computer science}, volume = {2143}, journal = {Lecture notes in computer science}, editor = {Benferhat, Salem and Besnard, Philippe}, publisher = {Springer}, address = {Berlin, Heidelberg}, isbn = {3-540-42464-4}, doi = {10.1007/3-540-44652-4}, pages = {818 S.}, year = {2001}, language = {en} } @book{Juergensen2006, author = {J{\"u}rgensen, Helmut}, title = {Synchronization}, series = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, volume = {2006, 5}, journal = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, publisher = {Univ.}, address = {Potsdam}, issn = {0946-7580}, pages = {21 S.}, year = {2006}, language = {en} } @article{EhlenzBergnerSchroeder2016, author = {Ehlenz, Matthias and Bergner, Nadine and Schroeder, Ulrik}, title = {Synergieeffekte zwischen Fach- und Lehramtsstudierenden in Softwarepraktika}, series = {Commentarii informaticae didacticae (CID)}, journal = {Commentarii informaticae didacticae (CID)}, number = {10}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, isbn = {978-3-86956-376-3}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-94875}, pages = {99 -- 102}, year = {2016}, abstract = {Dieser Beitrag diskutiert die Konzeption eines Software-Projektpraktikums im Bereich E-Learning, welches Lehramts- und Fachstudierenden der Informatik erm{\"o}glicht, voneinander zu profitieren und praxisrelevante Ergebnisse generiert. Vorbereitungen, Organisation und Durchf{\"u}hrung werden vorgestellt und diskutiert. Den Abschluss bildet ein Ausblick auf die Fortf{\"u}hrung des Konzepts und den Ausbau des Forschungsgebietes.}, language = {de} } @article{HartjeGoesselSogomonyan1997, author = {Hartje, Hendrik and G{\"o}ssel, Michael and Sogomonyan, Egor S.}, title = {Synthesis of code-disjoint combinational circuits}, year = {1997}, language = {en} } @article{MoschaninSaposhnikovSaposhnikovetal.1996, author = {Moschanin, Wladimir and Saposhnikov, Vl. V. and Saposhnikov, Va. V. and G{\"o}ssel, Michael}, title = {Synthesis of self-dual multi-output combinational circuits for on-line Teting}, year = {1996}, language = {en} } @article{GerbserSchaub2006, author = {Gerbser, Martin and Schaub, Torsten H.}, title = {Tableau calculi for answer set programming}, issn = {0302-9743}, doi = {10.1007/11799573}, year = {2006}, language = {en} } @article{GebserSchaub2013, author = {Gebser, Martin and Schaub, Torsten H.}, title = {Tableau calculi for logic programs under answer set semantics}, series = {ACM transactions on computational logic}, volume = {14}, journal = {ACM transactions on computational logic}, number = {2}, publisher = {Association for Computing Machinery}, address = {New York}, issn = {1529-3785}, doi = {10.1145/2480759.2480767}, pages = {40}, year = {2013}, abstract = {We introduce formal proof systems based on tableau methods for analyzing computations in Answer Set Programming (ASP). Our approach furnishes fine-grained instruments for characterizing operations as well as strategies of ASP solvers. The granularity is detailed enough to capture a variety of propagation and choice methods of algorithms used for ASP solving, also incorporating SAT-based and conflict-driven learning approaches to some extent. This provides us with a uniform setting for identifying and comparing fundamental properties of ASP solving approaches. In particular, we investigate their proof complexities and show that the run-times of best-case computations can vary exponentially between different existing ASP solvers. Apart from providing a framework for comparing ASP solving approaches, our characterizations also contribute to their understanding by pinning down the constitutive atomic operations. Furthermore, our framework is flexible enough to integrate new inference patterns, and so to study their relation to existing ones. To this end, we generalize our approach and provide an extensible basis aiming at a modular incorporation of additional language constructs. This is exemplified by augmenting our basic tableau methods with cardinality constraints and disjunctions.}, language = {en} } @article{LaiDavisEickelmannetal.2015, author = {Lai, Kwok-Wing and Davis, Niki and Eickelmann, Birgit and Erstad, Ola and Fisser, Petra and Gibson, David and Khaddage, Ferial and Knezek, Gerald and Webb, Mary}, title = {Tackling Educational Challenges in a Digitally Networked World}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82997}, pages = {415 -- 423}, year = {2015}, language = {en} } @phdthesis{Luebbe2011, author = {L{\"u}bbe, Alexander}, title = {Tangible business process modeling : design and evaluation of a process model elicitation Technique}, address = {Potsdam}, pages = {116 S.}, year = {2011}, language = {en} } @misc{PatilHaiderPopeetal.2011, author = {Patil, Kaustubh R. and Haider, Peter and Pope, Phillip B. and Turnbaugh, Peter J. and Morrison, Mark and Scheffer, Tobias and McHardy, Alice C.}, title = {Taxonomic metagenome sequence assignment with structured output models}, series = {Nature methods : techniques for life scientists and chemists}, volume = {8}, journal = {Nature methods : techniques for life scientists and chemists}, number = {3}, publisher = {Nature Publ. Group}, address = {London}, issn = {1548-7091}, doi = {10.1038/nmeth0311-191}, pages = {191 -- 192}, year = {2011}, language = {en} } @article{BarnesKennewell2015, author = {Barnes, Jan and Kennewell, Steve}, title = {Teacher Perceptions of Key Competencies in ICT}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82604}, pages = {61 -- 75}, year = {2015}, abstract = {Regardless of what is intended by government curriculum specifications and advised by educational experts, the competencies taught and learned in and out of classrooms can vary considerably. In this paper, we discuss in particular how we can investigate the perceptions that individual teachers have of competencies in ICT, and how these and other factors may influence students' learning. We report case study research which identifies contradictions within the teaching of ICT competencies as an activity system, highlighting issues concerning the object of the curriculum, the roles of the participants and the school cultures. In a particular case, contradictions in the learning objectives between higher order skills and the use of application tools have been resolved by a change in the teacher's perceptions which have not led to changes in other aspects of the activity system. We look forward to further investigation of the effects of these contradictions in other case studies and on forthcoming curriculum change.}, language = {en} } @article{GrillenbergerRomeike2015, author = {Grillenberger, Andreas and Romeike, Ralf}, title = {Teaching Data Management}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82648}, pages = {133 -- 150}, year = {2015}, abstract = {Data management is a central topic in computer science as well as in computer science education. Within the last years, this topic is changing tremendously, as its impact on daily life becomes increasingly visible. Nowadays, everyone not only needs to manage data of various kinds, but also continuously generates large amounts of data. In addition, Big Data and data analysis are intensively discussed in public dialogue because of their influences on society. For the understanding of such discussions and for being able to participate in them, fundamental knowledge on data management is necessary. Especially, being aware of the threats accompanying the ability to analyze large amounts of data in nearly real-time becomes increasingly important. This raises the question, which key competencies are necessary for daily dealings with data and data management. In this paper, we will first point out the importance of data management and of Big Data in daily life. On this basis, we will analyze which are the key competencies everyone needs concerning data management to be able to handle data in a proper way in daily life. Afterwards, we will discuss the impact of these changes in data management on computer science education and in particular database education.}, language = {en} } @article{Schiller2015, author = {Schiller, Thomas}, title = {Teaching Information Security (as Part of Key Competencies)}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82960}, pages = {401 -- 404}, year = {2015}, abstract = {The poster and abstract describe the importance of teaching information security in school. After a short description of information security and important aspects, I will show, how information security fits into different guidelines or models for computer science educations and that it is therefore on of the key competencies. Afterwards I will present you a rough insight of teaching information security in Austria.}, language = {en} } @article{BanbaraInoueKaufmannetal.2018, author = {Banbara, Mutsunori and Inoue, Katsumi and Kaufmann, Benjamin and Okimoto, Tenda and Schaub, Torsten H. and Soh, Takehide and Tamura, Naoyuki and Wanko, Philipp}, title = {teaspoon}, series = {Annals of operation research}, volume = {275}, journal = {Annals of operation research}, number = {1}, publisher = {Springer}, address = {Dordrecht}, issn = {0254-5330}, doi = {10.1007/s10479-018-2757-7}, pages = {3 -- 37}, year = {2018}, abstract = {Answer Set Programming (ASP) is an approach to declarative problem solving, combining a rich yet simple modeling language with high performance solving capacities. We here develop an ASP-based approach to curriculum-based course timetabling (CB-CTT), one of the most widely studied course timetabling problems. The resulting teaspoon system reads a CB-CTT instance of a standard input format and converts it into a set of ASP facts. In turn, these facts are combined with a first-order encoding for CB-CTT solving, which can subsequently be solved by any off-the-shelf ASP systems. We establish the competitiveness of our approach by empirically contrasting it to the best known bounds obtained so far via dedicated implementations. Furthermore, we extend the teaspoon system to multi-objective course timetabling and consider minimal perturbation problems.}, language = {en} } @inproceedings{KiyLucke2016, author = {Kiy, Alexander and Lucke, Ulrike}, title = {Technical Approaches for Personal Learning Environments}, series = {2016 IEEE 16th International Conference on Advanced Learning Technologies (ICALT)}, booktitle = {2016 IEEE 16th International Conference on Advanced Learning Technologies (ICALT)}, doi = {10.1109/icalt.2016.122}, year = {2016}, abstract = {The term Personal Learning Environment (PLE) is associated with the desire to put the learner in control of his own learning process, so that he is able to set and accomplish the desired learning goals at the right time with the learning environment chosen by him. Gradually, such a learning environment includes several digital content, services and tools. It is thus summarized as the Virtual Learning Environment (VLE). Even though the construction of an individual PLE is a complex task, several approaches to support this process already exist. They mostly occur under the umbrella term PLE or with little accentuations like iPLE, which especially live within the context of institutions. This paper sums up the variety of attempts and technical approaches to establish a PLE and suggests a categorization for them.}, language = {en} } @phdthesis{Maass2009, author = {Maaß, Stefan}, title = {Techniken zur automatisierten Annotation interaktiver geovirtueller 3D-Umgebungen}, address = {Potsdam}, pages = {111 S.}, year = {2009}, language = {de} } @incollection{KiyWoelfertLucke2016, author = {Kiy, Alexander and W{\"o}lfert, Volker and Lucke, Ulrike}, title = {Technische Unterst{\"u}tzung zur Durchf{\"u}hrung von Massenklausuren}, series = {DeLFI 2016 - Die 14. E-Learning Fachtagung Informatik, 11.-14. September 2016, Potsdam}, booktitle = {DeLFI 2016 - Die 14. E-Learning Fachtagung Informatik, 11.-14. September 2016, Potsdam}, number = {P-262}, editor = {Lucke, Ulrike and Schwill, Andreas and Zender, Raphael}, publisher = {Gesellschaft f{\"u}r Informatik e.V.}, address = {Bonn}, isbn = {978-3-88579-656-5}, publisher = {Universit{\"a}t Potsdam}, pages = {137 -- 148}, year = {2016}, abstract = {Nach wie vor stellen Klausuren mit hunderten Studierenden, sogenannte Massenklausuren, insbesondere Lehrende vor organisatorische Herausforderungen, die es jedes Semester aufs Neue zu bew{\"a}ltigen gilt. Dieser Beitrag zeigt technische M{\"o}glichkeiten zur Unterst{\"u}tzung und Durchf{\"u}hrung von Massenklausuren auf. Hierzu werden zun{\"a}chst verschiedene Verfahren der Klausurvorbereitung und -organisation aus unterschiedlichen Fachdisziplinen erfasst und analysiert. Im Anschluss wird ein verallgemeinerbarer Prozess konzipiert, der s{\"a}mtliche Schritte von der Planung, Erstellung, Durchf{\"u}hrung, Kontrolle bis hin zur Archivierung ber{\"u}cksichtigt. Zur Unterst{\"u}tzung der einzelnen Prozessschritte werden technische Systeme vorgestellt, die mit den Bed{\"u}rfnissen der Pr{\"u}fungsverantwortlichen abgestimmt sind. Die entstandenen Systeme werden vorgestellt und die durchgef{\"u}hrten Zielgruppentests reflektiert.}, language = {de} } @phdthesis{Gericke2014, author = {Gericke, Lutz}, title = {Tele-Board - Supporting and analyzing creative collaboration in synchronous and asynchronous scenario}, pages = {186}, year = {2014}, language = {en} } @article{GumiennyMeinelGerickeetal.2011, author = {Gumienny, Raja and Meinel, Christoph and Gericke, Lutz and Quasthoff, Matthias and LoBue, Peter and Willems, Christian}, title = {Tele-board : enabling efficient collaboration in digital design spaces across time and distance}, isbn = {978-3-642-13756-3}, year = {2011}, language = {en} } @article{GerickeGumiennyMeinel2012, author = {Gericke, Lutz and Gumienny, Raja and Meinel, Christoph}, title = {Tele-board : folow the traces of your design process history}, year = {2012}, language = {en} } @article{GumiennyGerickeWenzeletal.2012, author = {Gumienny, Raja and Gericke, Lutz and Wenzel, Matthias and Meinel, Christoph}, title = {Tele-board in use : applying aq digital whiteboard system in different situations and setups}, isbn = {978-3-642-31990-7}, year = {2012}, language = {en} } @misc{Frank2013, type = {Master Thesis}, author = {Frank, Mario}, title = {TEMPLAR : efficient determination of relevant axioms in big formula sets for theorem proving}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-72112}, school = {Universit{\"a}t Potsdam}, year = {2013}, abstract = {This document presents a formula selection system for classical first order theorem proving based on the relevance of formulae for the proof of a conjecture. It is based on unifiability of predicates and is also able to use a linguistic approach for the selection. The scope of the technique is the reduction of the set of formulae and the increase of the amount of provable conjectures in a given time. Since the technique generates a subset of the formula set, it can be used as a preprocessor for automated theorem proving. The document contains the conception, implementation and evaluation of both selection concepts. While the one concept generates a search graph over the negation normal forms or Skolem normal forms of the given formulae, the linguistic concept analyses the formulae and determines frequencies of lexemes and uses a tf-idf weighting algorithm to determine the relevance of the formulae. Though the concept is built for first order logic, it is not limited to it. The concept can be used for higher order and modal logik, too, with minimal adoptions. The system was also evaluated at the world championship of automated theorem provers (CADE ATP Systems Competition, CASC-24) in combination with the leanCoP theorem prover and the evaluation of the results of the CASC and the benchmarks with the problems of the CASC of the year 2012 (CASC-J6) show that the concept of the system has positive impact to the performance of automated theorem provers. Also, the benchmarks with two different theorem provers which use different calculi have shown that the selection is independent from the calculus. Moreover, the concept of TEMPLAR has shown to be competitive to some extent with the concept of SinE and even helped one of the theorem provers to solve problems that were not (or slower) solved with SinE selection in the CASC. Finally, the evaluation implies that the combination of the unification based and linguistic selection yields more improved results though no optimisation was done for the problems.}, language = {en} } @phdthesis{Huang2006, author = {Huang, Wanjun}, title = {Temporary binding for dynamic middleware construction and web services composition}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-7672}, school = {Universit{\"a}t Potsdam}, year = {2006}, abstract = {With increasing number of applications in Internet and mobile environments, distributed software systems are demanded to be more powerful and flexible, especially in terms of dynamism and security. This dissertation describes my work concerning three aspects: dynamic reconfiguration of component software, security control on middleware applications, and web services dynamic composition. Firstly, I proposed a technology named Routing Based Workflow (RBW) to model the execution and management of collaborative components and realize temporary binding for component instances. The temporary binding means component instances are temporarily loaded into a created execution environment to execute their functions, and then are released to their repository after executions. The temporary binding allows to create an idle execution environment for all collaborative components, on which the change operations can be immediately carried out. The changes on execution environment will result in a new collaboration of all involved components, and also greatly simplifies the classical issues arising from dynamic changes, such as consistency preserving etc. To demonstrate the feasibility of RBW, I created a dynamic secure middleware system - the Smart Data Server Version 3.0 (SDS3). In SDS3, an open source implementation of CORBA is adopted and modified as the communication infrastructure, and three secure components managed by RBW, are created to enhance the security on the access of deployed applications. SDS3 offers multi-level security control on its applications from strategy control to application-specific detail control. For the management by RBW, the strategy control of SDS3 applications could be dynamically changed by reorganizing the collaboration of the three secure components. In addition, I created the Dynamic Services Composer (DSC) based on Apache open source projects, Apache Axis and WSIF. In DSC, RBW is employed to model the interaction and collaboration of web services and to enable the dynamic changes on the flow structure of web services. Finally, overall performance tests were made to evaluate the efficiency of the developed RBW and SDS3. The results demonstrated that temporary binding of component instances makes slight impacts on the execution efficiency of components, and the blackout time arising from dynamic changes can be extremely reduced in any applications.}, subject = {Middleware}, language = {en} } @article{HollmannFrohmeEndrullatetal.2020, author = {Hollmann, Susanne and Frohme, Marcus and Endrullat, Christoph and Kremer, Andreas and D'Elia, Domenica and Regierer, Babette and Nechyporenko, Alina}, title = {Ten simple rules on how to write a standard operating procedure}, series = {PLOS Computational Biology}, volume = {16}, journal = {PLOS Computational Biology}, number = {9}, publisher = {PLOS}, address = {San Francisco}, pages = {10}, year = {2020}, abstract = {Research publications and data nowadays should be publicly available on the internet and, theoretically, usable for everyone to develop further research, products, or services. The long-term accessibility of research data is, therefore, fundamental in the economy of the research production process. However, the availability of data is not sufficient by itself, but also their quality must be verifiable. Measures to ensure reuse and reproducibility need to include the entire research life cycle, from the experimental design to the generation of data, quality control, statistical analysis, interpretation, and validation of the results. Hence, high-quality records, particularly for providing a string of documents for the verifiable origin of data, are essential elements that can act as a certificate for potential users (customers). These records also improve the traceability and transparency of data and processes, therefore, improving the reliability of results. Standards for data acquisition, analysis, and documentation have been fostered in the last decade driven by grassroot initiatives of researchers and organizations such as the Research Data Alliance (RDA). Nevertheless, what is still largely missing in the life science academic research are agreed procedures for complex routine research workflows. Here, well-crafted documentation like standard operating procedures (SOPs) offer clear direction and instructions specifically designed to avoid deviations as an absolute necessity for reproducibility. Therefore, this paper provides a standardized workflow that explains step by step how to write an SOP to be used as a starting point for appropriate research documentation.}, language = {en} } @misc{HollmannFrohmeEndrullatetal.2020, author = {Hollmann, Susanne and Frohme, Marcus and Endrullat, Christoph and Kremer, Andreas and D'Elia, Domenica and Regierer, Babette and Nechyporenko, Alina}, title = {Ten simple rules on how to write a standard operating procedure}, series = {Postprints der Universit{\"a}t Potsdam : Mathematisch-Naturwissenschaftliche Reihe}, journal = {Postprints der Universit{\"a}t Potsdam : Mathematisch-Naturwissenschaftliche Reihe}, number = {9}, issn = {1866-8372}, doi = {10.25932/publishup-52587}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-525877}, pages = {12}, year = {2020}, abstract = {Research publications and data nowadays should be publicly available on the internet and, theoretically, usable for everyone to develop further research, products, or services. The long-term accessibility of research data is, therefore, fundamental in the economy of the research production process. However, the availability of data is not sufficient by itself, but also their quality must be verifiable. Measures to ensure reuse and reproducibility need to include the entire research life cycle, from the experimental design to the generation of data, quality control, statistical analysis, interpretation, and validation of the results. Hence, high-quality records, particularly for providing a string of documents for the verifiable origin of data, are essential elements that can act as a certificate for potential users (customers). These records also improve the traceability and transparency of data and processes, therefore, improving the reliability of results. Standards for data acquisition, analysis, and documentation have been fostered in the last decade driven by grassroot initiatives of researchers and organizations such as the Research Data Alliance (RDA). Nevertheless, what is still largely missing in the life science academic research are agreed procedures for complex routine research workflows. Here, well-crafted documentation like standard operating procedures (SOPs) offer clear direction and instructions specifically designed to avoid deviations as an absolute necessity for reproducibility. Therefore, this paper provides a standardized workflow that explains step by step how to write an SOP to be used as a starting point for appropriate research documentation.}, language = {en} } @article{SinghSogomonyanGoesseletal.1999, author = {Singh, Adit D. and Sogomonyan, Egor S. and G{\"o}ssel, Michael and Seuring, Markus}, title = {Testability evaluation of sequential designs incorporating the multi-mode scannable memory element}, year = {1999}, language = {en} } @phdthesis{Hilscher2010, author = {Hilscher, Martin}, title = {Testdatenkompaktion durch beschleunigte Schieberegister mit vielen X-Werten}, address = {Potsdam}, pages = {112, 5 S.}, year = {2010}, language = {de} } @phdthesis{Kunz1996, author = {Kunz, Wolfgang}, title = {Testing techniques in logic synthesis}, pages = {189 S.}, year = {1996}, language = {en} } @phdthesis{Lorenz2011, author = {Lorenz, Haik}, title = {Texturierung und Visualisierung virtueller 3D-Stadtmodelle}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-53879}, school = {Universit{\"a}t Potsdam}, year = {2011}, abstract = {Im Mittelpunkt dieser Arbeit stehen virtuelle 3D-Stadtmodelle, die Objekte, Ph{\"a}nomene und Prozesse in urbanen R{\"a}umen in digitaler Form repr{\"a}sentieren. Sie haben sich zu einem Kernthema von Geoinformationssystemen entwickelt und bilden einen zentralen Bestandteil geovirtueller 3D-Welten. Virtuelle 3D-Stadtmodelle finden nicht nur Verwendung als Mittel f{\"u}r Experten in Bereichen wie Stadtplanung, Funknetzplanung, oder L{\"a}rmanalyse, sondern auch f{\"u}r allgemeine Nutzer, die realit{\"a}tsnah dargestellte virtuelle St{\"a}dte in Bereichen wie B{\"u}rgerbeteiligung, Tourismus oder Unterhaltung nutzen und z. B. in Anwendungen wie GoogleEarth eine r{\"a}umliche Umgebung intuitiv erkunden und durch eigene 3D-Modelle oder zus{\"a}tzliche Informationen erweitern. Die Erzeugung und Darstellung virtueller 3D-Stadtmodelle besteht aus einer Vielzahl von Prozessschritten, von denen in der vorliegenden Arbeit zwei n{\"a}her betrachtet werden: Texturierung und Visualisierung. Im Bereich der Texturierung werden Konzepte und Verfahren zur automatischen Ableitung von Fototexturen aus georeferenzierten Schr{\"a}gluftbildern sowie zur Speicherung oberfl{\"a}chengebundener Daten in virtuellen 3D-Stadtmodellen entwickelt. Im Bereich der Visualisierung werden Konzepte und Verfahren f{\"u}r die multiperspektivische Darstellung sowie f{\"u}r die hochqualitative Darstellung nichtlinearer Projektionen virtueller 3D-Stadtmodelle in interaktiven Systemen vorgestellt. Die automatische Ableitung von Fototexturen aus georeferenzierten Schr{\"a}gluftbildern erm{\"o}glicht die Veredelung vorliegender virtueller 3D-Stadtmodelle. Schr{\"a}gluftbilder bieten sich zur Texturierung an, da sie einen Großteil der Oberfl{\"a}chen einer Stadt, insbesondere Geb{\"a}udefassaden, mit hoher Redundanz erfassen. Das Verfahren extrahiert aus dem verf{\"u}gbaren Bildmaterial alle Ansichten einer Oberfl{\"a}che und f{\"u}gt diese pixelpr{\"a}zise zu einer Textur zusammen. Durch Anwendung auf alle Oberfl{\"a}chen wird das virtuelle 3D-Stadtmodell fl{\"a}chendeckend texturiert. Der beschriebene Ansatz wurde am Beispiel des offiziellen Berliner 3D-Stadtmodells sowie der in GoogleEarth integrierten Innenstadt von M{\"u}nchen erprobt. Die Speicherung oberfl{\"a}chengebundener Daten, zu denen auch Texturen z{\"a}hlen, wurde im Kontext von CityGML, einem international standardisierten Datenmodell und Austauschformat f{\"u}r virtuelle 3D-Stadtmodelle, untersucht. Es wird ein Datenmodell auf Basis computergrafischer Konzepte entworfen und in den CityGML-Standard integriert. Dieses Datenmodell richtet sich dabei an praktischen Anwendungsf{\"a}llen aus und l{\"a}sst sich dom{\"a}nen{\"u}bergreifend verwenden. Die interaktive multiperspektivische Darstellung virtueller 3D-Stadtmodelle erg{\"a}nzt die gewohnte perspektivische Darstellung nahtlos um eine zweite Perspektive mit dem Ziel, den Informationsgehalt der Darstellung zu erh{\"o}hen. Diese Art der Darstellung ist durch die Panoramakarten von H. C. Berann inspiriert; Hauptproblem ist die {\"U}bertragung des multiperspektivischen Prinzips auf ein interaktives System. Die Arbeit stellt eine technische Umsetzung dieser Darstellung f{\"u}r 3D-Grafikhardware vor und demonstriert die Erweiterung von Vogel- und Fußg{\"a}ngerperspektive. Die hochqualitative Darstellung nichtlinearer Projektionen beschreibt deren Umsetzung auf 3D-Grafikhardware, wobei neben der Bildwiederholrate die Bildqualit{\"a}t das wesentliche Entwicklungskriterium ist. Insbesondere erlauben die beiden vorgestellten Verfahren, dynamische Geometrieverfeinerung und st{\"u}ckweise perspektivische Projektionen, die uneingeschr{\"a}nkte Nutzung aller hardwareseitig verf{\"u}gbaren, qualit{\"a}tssteigernden Funktionen wie z.~B. Bildraumgradienten oder anisotroper Texturfilterung. Beide Verfahren sind generisch und unterst{\"u}tzen verschiedene Projektionstypen. Sie erm{\"o}glichen die anpassungsfreie Verwendung g{\"a}ngiger computergrafischer Effekte wie Stilisierungsverfahren oder prozeduraler Texturen f{\"u}r nichtlineare Projektionen bei optimaler Bildqualit{\"a}t. Die vorliegende Arbeit beschreibt wesentliche Technologien f{\"u}r die Verarbeitung virtueller 3D-Stadtmodelle: Zum einen lassen sich mit den Ergebnissen der Arbeit Texturen f{\"u}r virtuelle 3D-Stadtmodelle automatisiert herstellen und als eigenst{\"a}ndige Attribute in das virtuelle 3D-Stadtmodell einf{\"u}gen. Somit tr{\"a}gt diese Arbeit dazu bei, die Herstellung und Fortf{\"u}hrung texturierter virtueller 3D-Stadtmodelle zu verbessern. Zum anderen zeigt die Arbeit Varianten und technische L{\"o}sungen f{\"u}r neuartige Projektionstypen f{\"u}r virtueller 3D-Stadtmodelle in interaktiven Visualisierungen. Solche nichtlinearen Projektionen stellen Schl{\"u}sselbausteine dar, um neuartige Benutzungsschnittstellen f{\"u}r und Interaktionsformen mit virtuellen 3D-Stadtmodellen zu erm{\"o}glichen, insbesondere f{\"u}r mobile Ger{\"a}te und immersive Umgebungen.}, language = {de} } @book{Schaub1999, author = {Schaub, Torsten H.}, title = {The automation of reasoning with incomplete information : from semantic foundations to efficient computation}, series = {Lecture notes in computer science}, volume = {1409}, journal = {Lecture notes in computer science}, publisher = {Springer}, address = {Berlin}, isbn = {3-540-64515-2}, doi = {10.1007/BFb0054963}, pages = {XI, 159 S.}, year = {1999}, language = {en} } @article{BlankertzMuellerCurioetal.2004, author = {Blankertz, Benjamin and M{\"u}ller, Klaus-Robert and Curio, Gabriel and Vaughan, Theresa M. and Schalk, Gerwin and Wolpaw, Jonathan R. and Schlogl, Alois and Neuper, Christa and Pfurtscheller, Gert and Hinterberger, Thilo and Schroder, Michael and Birbaumer, Niels}, title = {The BCI competition 2003 : Progress and perspectives in detection and discrimination of EEG single trials}, issn = {0018-9294}, year = {2004}, abstract = {Interest in developing a new method of man-to-machine communication-a brain-computer interface (BCI)-has grown steadily over the past few decades. BCIs create a new communication channel between the brain and an output device by bypassing conventional motor output pathways of nerves and muscles. These systems use signals recorded from the scalp, the surface of the cortex, or from inside the brain to enable users to control a variety of applications including simple word-processing software and orthotics. BCI technology could therefore provide a new communication and control option for individuals who cannot otherwise express their wishes to the outside world. Signal processing and classification methods are essential tools in the development of improved BCI technology. We organized the BCI Competition 2003 to evaluate the current state of the art of these tools. Four laboratories well versed in EEG-based BCI research provided six data sets in a documented format. We made these data sets (i.e., labeled training sets and unlabeled test sets) and their descriptions available on the Internet. The goal in the competition was to maximize the performance measure for the test labels. Researchers worldwide tested their algorithms and competed for the best classification results. This paper describes the six data sets and the results and function of the most successful algorithms}, language = {en} } @article{BlankertzMuellerKrusienskietal.2006, author = {Blankertz, Benjamin and M{\"u}ller, Klaus-Robert and Krusienski, Dean and Schalk, Gerwin and Wolpaw, Jonathan R. and Schl{\"o}gl, Alois and Pfurtscheller, Gert and Millan, Jos{\´e} del R. and Schr{\"o}der, Michael and Birbaumer, Niels}, title = {The BCI competition III : validating alternative approaches to actual BCI problems}, issn = {1534-4320}, doi = {10.1109/Tnsre.2006.875642}, year = {2006}, abstract = {A brain-computer interface (BCI) is a system that allows its users to control external devices with brain activity. Although the proof-of-concept was given decades ago, the reliable translation of user intent into device control commands is still a major challenge. Success requires the effective interaction of two adaptive controllers: the user's brain, which produces brain activity that encodes intent, and the BCI system, which translates that activity into device control commands. In order to facilitate this interaction, many laboratories are exploring a variety of signal analysis techniques to improve the adaptation of the BCI system to the user. In the literature, many machine learning and pattern classification algorithms have been reported to give impressive results when applied to BCI data in offline analyses. However, it is more difficult to evaluate their relative value for actual online use. BCI data competitions have been organized to provide objective formal evaluations of alternative methods. Prompted by the great interest in the first two BCI Competitions, we organized the third BCI Competition to address several of the most difficult and important analysis problems in BCI research. The paper describes the data sets that were provided to the competitors and gives an overview of the results.}, language = {en} } @phdthesis{Schneidenbach2009, author = {Schneidenbach, Lars}, title = {The benefits of one-sided communication interfaces for cluster computing}, address = {Potsdam}, pages = {208 S. : graph. Darst.}, year = {2009}, language = {en} } @article{BlankertzDornhegeKrauledatetal.2006, author = {Blankertz, Benjamin and Dornhege, Guido and Krauledat, Matthias and M{\"u}ller, Klaus-Robert and Kunzmann, Volker and Losch, Florian and Curio, Gabriel}, title = {The Berlin brain-computer interface : EEG-based communication without subject training}, issn = {1534-4320}, doi = {10.1109/Tnsre.2006.875557}, year = {2006}, abstract = {The Berlin Brain-Computer Interface (BBCI) project develops a noninvasive BCI system whose key features are 1) the use of well-established motor competences as control paradigms, 2) high-dimensional features from 128-channel electroencephalogram (EEG), and 3) advanced machine learning techniques. As reported earlier, our experiments demonstrate that very high information transfer rates can be achieved using the readiness potential (RP) when predicting the laterality of upcoming left-versus right-hand movements in healthy subjects. A more recent study showed that the RP similarily accompanies phantom movements in arm amputees, but the signal strength decreases with longer loss of the limb. In a complementary approach, oscillatory features are used to discriminate imagined movements (left hand versus right hand versus foot). In a recent feedback study with six healthy subjects with no or very little experience with BCI control, three subjects achieved an information transfer rate above 35 bits per minute (bpm), and further two subjects above 24 and 15 bpm, while one subject could not achieve any BCI control. These results are encouraging for an EEG-based BCI system in untrained subjects that is independent of peripheral nervous system activity and does not rely on evoked potentials even when compared to results with very well-trained subjects operating other BCI systems}, language = {en} } @article{HawroPrzybylowiczSpindleretal.2021, author = {Hawro, Tomasz and Przybylowicz, Katarzyna and Spindler, Max and Hawro, Marlena and Steć, Michał and Altrichter, Sabine and Weller, Karsten and Magerl, Markus and Reidel, Ulrich and Alarbeed, Ezzat and Alraboni, Ola and Maurer, Marcus and Metz, Martin}, title = {The characteristics and impact of pruritus in adult dermatology patients}, series = {Journal of the American Academy of Dermatology}, volume = {84}, journal = {Journal of the American Academy of Dermatology}, number = {3}, publisher = {Elsevier}, address = {Amsterdam [u.a.]}, issn = {0190-9622}, doi = {10.1016/J.JAAD.2020.08.035}, pages = {691 -- 700}, year = {2021}, abstract = {Background: Pruritus often accompanies chronic skin diseases, exerting considerable burden on many areas of patient functioning; this burden and the features of pruritus remain insufficiently characterized. Objective: To investigate characteristics, including localization patterns, and burden of pruritus in patients with chronic dermatoses. Methods: We recruited 800 patients with active chronic skin diseases. We assessed pruritus intensity, localization, and further characteristics. We used validated questionnaires to assess quality of life, work productivity and activity impairment, anxiety, depression, and sleep quality. Results: Nine out of every 10 patients had experienced pruritus throughout their disease and 73\% in the last 7 days. Pruritus often affected the entire body and was not restricted to skin lesions. Patients with moderate to severe pruritus reported significantly more impairment to their sleep quality and work productivity, and they were more depressed and anxious than control individuals and patients with mild or no pruritus. Suicidal ideations were highly prevalent in patients with chronic pruritus (18.5\%) and atopic dermatitis (11.8\%). Conclusions: Pruritus prevalence and intensity are very high across all dermatoses studied; intensity is linked to impairment in many areas of daily functioning. Effective treatment strategies are urgently required to treat pruritus and the underlying skin disease. ( J Am Acad Dermatol 2021;84:691-700.)}, language = {en} } @article{ThienenNoweskiMeineletal.2011, author = {Thienen, Julia von and Noweski, Christine and Meinel, Christoph and Rauth, Ingo}, title = {The co-evolution of theory and practice in design thinking - or - "Mind the oddness trap!"}, isbn = {978-3-642-13756-3}, year = {2011}, language = {en} } @article{BoernerBulatovChenetal.2009, author = {Boerner, Ferdinand and Bulatov, Andrei and Chen, Hubie and Jeavons, Peter and Krokhin, Andrei}, title = {The complexity of constraint satisfaction games and QCSP}, issn = {0890-5401}, doi = {10.1016/J.Ic.2009.05.003}, year = {2009}, abstract = {We study the complexity of two-person constraint satisfaction games. An instance of such a game is given by a collection of constraints on overlapping sets of variables, and the two players alternately make moves assigning values from a finite domain to the variables, in a specified order. The first player tries to satisfy all constraints, while the other tries to break at least one constraint: the goal is to decide whether the first player has a winning strategy. We show that such games can be conveniently represented by a logical form of quantified constraint satisfaction, where an instance is given by a first-order sentence in which quantifiers alternate and the quantifier-free part is a conjunction of (positive) atomic formulas; the goal is to decide whether the sentence is true. While the problem of deciding such a game is PSPACE-complete in general, by restricting the set of allowed constraint predicates, one can obtain infinite classes of constraint satisfaction games of lower complexity. We use the quantified constraint satisfaction framework to study how the complexity of deciding such a game depends on the parameter set of allowed predicates. With every predicate. one can associate certain predicate-preserving operations, called polymorphisms. We show that the complexity of our games is determined by the surjective polymorphisms of the constraint predicates. We illustrate how this result can be used by identifying the complexity of a wide variety of constraint satisfaction games.}, language = {en} } @article{LamprechtWickert2014, author = {Lamprecht, Anna-Lena and Wickert, Alexander}, title = {The Course's SIB Libraries}, series = {Process Design for Natural Scientists: an agile model-driven approach}, journal = {Process Design for Natural Scientists: an agile model-driven approach}, number = {500}, editor = {Lambrecht, Anna-Lena and Margaria, Tiziana}, publisher = {Springer Verlag}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {30 -- 44}, year = {2014}, abstract = {This chapter gives a detailed description of the service framework underlying all the example projects that form the foundation of this book. It describes the different SIB libraries that we made available for the course "Process modeling in the natural sciences" to provide the functionality that was required for the envisaged applications. The students used these SIB libraries to realize their projects.}, language = {en} } @phdthesis{Felgentreff2017, author = {Felgentreff, Tim}, title = {The Design and Implementation of Object-Constraint Programming}, school = {Universit{\"a}t Potsdam}, pages = {183}, year = {2017}, language = {en} } @article{JobstKoeppenLindbergetal.2012, author = {Jobst, Birgit and K{\"o}ppen, Eva and Lindberg, Tilmann and Moritz, Josephine and Rhinow, Holger and Meinel, Christoph}, title = {The faith-factor in design thinking : creative confidence through education at the design thinking schools Potsdam and Standford?}, isbn = {978-3-642-31990-7}, year = {2012}, language = {en} } @article{Schaub1998, author = {Schaub, Torsten H.}, title = {The family of default logics}, year = {1998}, language = {en} } @article{MeinelKlotz2006, author = {Meinel, Christoph and Klotz, Volker}, title = {The first 10 years of the ECCC digital library}, doi = {10.1145/1107458.1107484}, year = {2006}, language = {en} } @article{GebserLiuNamasivayametal.2007, author = {Gebser, Martin and Liu, Lengning and Namasivayam, Gayathri and Neumann, Andr{\´e} and Schaub, Torsten H. and Truszczynski, Miroslaw}, title = {The first answer set programming system competition}, isbn = {978-3-540- 72199-4}, year = {2007}, language = {en} } @article{MoshaninOtscheretnijDimitriev1998, author = {Moshanin, Vl. and Otscheretnij, Vitalij and Dimitriev, Alexej}, title = {The impact of logic optimization on concurrent error detection}, year = {1998}, language = {en} } @book{PlattnerLeukert2015, author = {Plattner, Hasso and Leukert, Bernd}, title = {The in-memory revolution}, publisher = {Springer}, address = {Cham}, isbn = {978-3-319-16672-8}, pages = {275}, year = {2015}, abstract = {This book describes the next generation of business applications enabled by SAP's in-memory database, SAP HANA. In particular, the authors show the substantial changes introduced in S4/HANA by switching to SAP HANA. Using numerous examples and use cases from the authors' wealth of real-world experience, it illustrates the quantum leap in performance made possible by the new technology. The book is written by two of the most prominent actors in the area of business application systems: Hasso Plattner, co-founder of SAP and inaugurator of the Hasso Plattner Institute at the University of Potsdam, and Bernd Leukert, member of the Executive Board and the Global Managing Board of SAP. This clearly structured, highly illustrated book takes an exciting new technology and presents the practicality and success of first mover applications.}, language = {en} } @book{OPUS4-27228, title = {The journal of supercomputing : trends in parallel operating systems}, editor = {Schr{\"o}der-Preikschat, Wolfgang and Wu, Min-You}, publisher = {Kluwer}, address = {Boston}, pages = {157 S.}, year = {1995}, language = {en} } @article{Saito2015, author = {Saito, Toshinori}, title = {The Key Competencies in Informatics and ICT viewed from Nussbaum's Ten Central Capabilities}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82718}, pages = {253 -- 266}, year = {2015}, abstract = {This article shows a discussion about the key competencies in informatics and ICT viewed from a philosophical foundation presented by Martha Nussbaum, which is known as 'ten central capabilities'. Firstly, the outline of 'The Capability Approach', which has been presented by Amartya Sen and Nussbaum as a theoretical framework of assessing the state of social welfare, will be explained. Secondly, the body of Nussbaum's ten central capabilities and the reason for being applied as the basis of discussion will be shown. Thirdly, the relationship between the concept of 'capability' and 'competency' is to be discussed. After that, the author's assumption of the key competencies in informatics and ICT led from the examination of Nussbaum's ten capabilities will be presented.}, language = {en} } @book{SchroederPreikschat1994, author = {Schr{\"o}der-Preikschat, Wolfgang}, title = {The logical design of parallel operating systems}, publisher = {Prentice Hall International}, address = {Englewood Cliffs}, pages = {370 S.}, year = {1994}, language = {en} } @article{SchulmeisterScheller1996, author = {Schulmeister, Thomas and Scheller, Frieder W.}, title = {The mathematics of exponential signal amplification in amperometric three enzyme electrodes}, year = {1996}, language = {en} } @book{Tepoyan2008, author = {Tepoyan, Liparit}, title = {The Mixed problem for a degenerate operator equation}, series = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Mathematik, Arbeitsgruppe Partiell}, journal = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Mathematik, Arbeitsgruppe Partiell}, publisher = {Univ.}, address = {Potsdam}, issn = {1437-739X}, pages = {13 S.}, year = {2008}, language = {en} } @book{Tepoyan2004, author = {Tepoyan, Liparit}, title = {The Neumann problem for a degenerate operator equation}, series = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Mathematik, Arbeitsgruppe Partiell}, journal = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Mathematik, Arbeitsgruppe Partiell}, publisher = {Univ.}, address = {Potsdam}, issn = {1437-739X}, pages = {11 S.}, year = {2004}, language = {en} } @article{SchroederPreikschatGiloi1995, author = {Schr{\"o}der-Preikschat, Wolfgang and Giloi, Wolfgang K.}, title = {The next generation parallel architecture - multiple executing threads}, year = {1995}, language = {en} } @article{AngerGebserLinkeetal.2005, author = {Anger, Christian and Gebser, Martin and Linke, Thomas and Neumann, Andre and Schaub, Torsten H.}, title = {The nomore++ approach to answer set solving}, year = {2005}, language = {en} } @article{AngerGebserLinkeetal.2005, author = {Anger, Christian and Gebser, Martin and Linke, Thomas and Neumann, Andre and Schaub, Torsten H.}, title = {The nomore++ approach to answer set solving}, year = {2005}, language = {en} } @article{GebserKaminskiKaufmannetal.2018, author = {Gebser, Martin and Kaminski, Roland and Kaufmann, Benjamin and L{\"u}hne, Patrick and Obermeier, Philipp and Ostrowski, Max and Romero Davila, Javier and Schaub, Torsten H. and Schellhorn, Sebastian and Wanko, Philipp}, title = {The Potsdam Answer Set Solving Collection 5.0}, series = {K{\"u}nstliche Intelligenz}, volume = {32}, journal = {K{\"u}nstliche Intelligenz}, number = {2-3}, publisher = {Springer}, address = {Heidelberg}, issn = {0933-1875}, doi = {10.1007/s13218-018-0528-x}, pages = {181 -- 182}, year = {2018}, abstract = {The Potsdam answer set solving collection, or Potassco for short, bundles various tools implementing and/or applying answer set programming. The article at hand succeeds an earlier description of the Potassco project published in Gebser et al. (AI Commun 24(2):107-124, 2011). Hence, we concentrate in what follows on the major features of the most recent, fifth generation of the ASP system clingo and highlight some recent resulting application systems.}, language = {en} } @article{DelgrandeSchaub2000, author = {Delgrande, James Patrick and Schaub, Torsten H.}, title = {The role of default logic in knowledge representation}, isbn = {0-7923-7224-7}, year = {2000}, language = {en} } @article{GebserMarateaRicca2020, author = {Gebser, Martin and Maratea, Marco and Ricca, Francesco}, title = {The Seventh Answer Set Programming Competition}, series = {Theory and practice of logic programming}, volume = {20}, journal = {Theory and practice of logic programming}, number = {2}, publisher = {Cambridge Univ. Press}, address = {Cambridge [u.a.]}, issn = {1471-0684}, doi = {10.1017/S1471068419000061}, pages = {176 -- 204}, year = {2020}, abstract = {Answer Set Programming (ASP) is a prominent knowledge representation language with roots in logic programming and non-monotonic reasoning. Biennial ASP competitions are organized in order to furnish challenging benchmark collections and assess the advancement of the state of the art in ASP solving. In this paper, we report on the design and results of the Seventh ASP Competition, jointly organized by the University of Calabria (Italy), the University of Genova (Italy), and the University of Potsdam (Germany), in affiliation with the 14th International Conference on Logic Programming and Non-Monotonic Reasoning (LPNMR 2017).}, language = {en} } @article{GebserMarateaRicca2017, author = {Gebser, Martin and Maratea, Marco and Ricca, Francesco}, title = {The sixth answer set programming competition}, series = {Journal of artificial intelligence research : JAIR}, volume = {60}, journal = {Journal of artificial intelligence research : JAIR}, publisher = {AI Access Found.}, address = {Marina del Rey}, issn = {1076-9757}, doi = {10.1613/jair.5373}, pages = {41 -- 95}, year = {2017}, abstract = {Answer Set Programming (ASP) is a well-known paradigm of declarative programming with roots in logic programming and non-monotonic reasoning. Similar to other closely related problemsolving technologies, such as SAT/SMT, QBF, Planning and Scheduling, advancements in ASP solving are assessed in competition events. In this paper, we report about the design and results of the Sixth ASP Competition, which was jointly organized by the University of Calabria (Italy), Aalto University (Finland), and the University of Genoa (Italy), in affiliation with the 13th International Conference on Logic Programming and Non-Monotonic Reasoning. This edition maintained some of the design decisions introduced in 2014, e.g., the conception of sub-tracks, the scoring scheme,and the adherence to a fixed modeling language in order to push the adoption of the ASP-Core-2 standard. On the other hand, it featured also some novelties, like a benchmark selection stage classifying instances according to their empirical hardness, and a "Marathon" track where the topperforming systems are given more time for solving hard benchmarks.}, language = {en} } @article{Haugsbakken2015, author = {Haugsbakken, Halvdan}, title = {The Student Learning Ecology}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82659}, pages = {151 -- 169}, year = {2015}, abstract = {Educational research on social media has showed that students use it for socialisation, personal communication, and informal learning. Recent studies have argued that students to some degree use social media to carry out formal schoolwork. This article gives an explorative account on how a small sample of Norwegian high school students use social media to self-organise formal schoolwork. This user pattern can be called a "student learning ecology", which is a user perspective on how participating students gain access to learning resources.}, language = {en} } @article{ChristensenKnezek2015, author = {Christensen, Rhonda and Knezek, Gerald}, title = {The Technology Proficiency Self-Assessment Questionnaire (TPSA)}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82838}, pages = {311 -- 318}, year = {2015}, abstract = {The Technology Proficiency Self-Assessment (TPSA) questionnaire has been used for 15 years in the USA and other nations as a self-efficacy measure for proficiencies fundamental to effective technology integration in the classroom learning environment. Internal consistency reliabilities for each of the five-item scales have typically ranged from .73 to .88 for preservice or inservice technology-using teachers. Due to changing technologies used in education, researchers sought to renovate partially obsolete items and extend self-efficacy assessment to new areas, such as social media and mobile learning. Analysis of 2014 data gathered on a new, 34 item version of the TPSA indicates that the four established areas of email, World Wide Web (WWW), integrated applications, and teaching with technology continue to form consistent scales with reliabilities ranging from .81 to .93, while the 14 new items gathered to represent emerging technologies and media separate into two scales, each with internal consistency reliabilities greater than .9. The renovated TPSA is deemed to be worthy of continued use in the teaching with technology context.}, language = {en} } @phdthesis{Schuenemann2011, author = {Sch{\"u}nemann, Bj{\"o}rn}, title = {The V2X simulation runtime infrastructure: VSimRTI}, address = {Potsdam}, pages = {163 S.}, year = {2011}, language = {en} } @article{NicolasSchaub1998, author = {Nicolas, Pascal and Schaub, Torsten H.}, title = {The XRay system : an implementation platform for local query-answering in default logics}, isbn = {3-540-65312-0}, year = {1998}, language = {en} } @article{Rolf2010, author = {Rolf, Arno}, title = {Themeng{\"a}rten in der Informatik-Ausbildung}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {4}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-64281}, pages = {7 -- 12}, year = {2010}, abstract = {Die M{\"o}glichkeiten sich zu informieren, am Leben der vielen Anderen teilzunehmen ist durch das Internet mit seinen Tweets, Google-Angeboten und sozialen Netzwerken wie Facebook ins Unermessliche gewachsen. Zugleich f{\"u}hlen sich viele Nutzer {\"u}berfordert und meinen, im Meer der Informationen zu ertrinken. So bekennt Frank Schirrmacher in seinem Buch Payback, dass er den geistigen Anforderungen unserer Zeit nicht mehr gewachsen ist. Sein Kopf komme nicht mehr mit. Er sei unkonzentriert, vergesslich und st{\"a}ndig abgelenkt. Das, was vielen zum Problem geworden ist, sehen viele Studierende eher pragmatisch. Der Wissenserwerb in Zeiten von Internet und E-Learning l{\"a}uft an Hochschulen h{\"a}ufig nach der Helene-Hegemann-Methode ab: Zun{\"a}chst machen sich die Studierenden, z.B. im Rahmen einer Studien- oder Hausarbeit, bei Wikipedia „schlau", ein Einstieg ist geschafft. Anschließend wird dieses Wissen mit Google angereichert. Damit ist {\"U}berblickswissen vorhanden. Mit geschickter copy-and-paste-Komposition l{\"a}sst sich daraus schon ein „Werk" erstellen. Der ein oder andere Studierende gibt sich mit diesem Wissenserwerb zufrieden und bricht seinen Lernprozess hier bereits ab. Nun ist zwar am Ende jeder Studierende f{\"u}r seinen Wissenserwerb selbst verantwortlich. Die erkennbar unbefriedigende Situation sollte die Hochschulen aber herausfordern, das Internet in Vorlesungen und Seminaren auszuprobieren und sinnvolle Anwendungen zu entwickeln. Beispiele gibt es durchaus. Unter der Metapher E-Learning hat sich ein umfangreicher Forschungsschwerpunkt an den Universit{\"a}ten entwickelt. Einige Beispiele von vielen: So hat der Osnabr{\"u}cker Informatik-Professor Oliver Vornberger seine Vorlesungen als Video ins Netz gestellt. Per RSS ist es m{\"o}glich, Sequenzen aufs iPod zu laden. Die {\"u}bliche Dozentenangst, dann w{\"u}rden sie ja vor leeren B{\"a}nken sitzen, scheint unbegr{\"u}ndet. Sie werden von den Studierenden vor allem zur Pr{\"u}fungsvorbereitung genutzt. Wie ist das Internet, das f{\"u}r die junge Generation zu einem alles andere verdr{\"a}ngenden Universalmedium geworden ist, didaktisch in die Hochschullehre einzubinden? Wie also ist konkret mit diesen Herausforderungen umzugehen? Dies soll uns im Folgenden besch{\"a}ftigen.}, language = {de} } @book{Wilhelm1997, author = {Wilhelm, Otto}, title = {Theoretische Grundlagen der Informatik : ein vorlesungsbegleitendes Manuskript f{\"u}r den internen Gebrauch am Institut ; Bd. 1 Automaten und Umwelten}, editor = {Budach, Lothar}, publisher = {Univ.}, address = {Potsdam}, pages = {54 S.}, year = {1997}, language = {de} } @article{SysłoKwiatkowska2015, author = {Sysło, Maciej M. and Kwiatkowska, Anna Beata}, title = {Think logarithmically!}, series = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, journal = {KEYCIT 2014 - Key Competencies in Informatics and ICT}, number = {7}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-82923}, pages = {371 -- 380}, year = {2015}, abstract = {We discuss here a number of algorithmic topics which we use in our teaching and in learning of mathematics and informatics to illustrate and document the power of logarithm in designing very efficient algorithms and computations - logarithmic thinking is one of the most important key competencies for solving real world practical problems. We demonstrate also how to introduce logarithm independently of mathematical formalism using a conceptual model for reducing a problem size by at least half. It is quite surprising that the idea, which leads to logarithm, is present in Euclid's algorithm described almost 2000 years before John Napier invented logarithm.}, language = {en} } @article{FloeterNicolasSchaubetal.2004, author = {Fl{\"o}ter, Andr{\´e} and Nicolas, Jacques and Schaub, Torsten H. and Selbig, Joachim}, title = {Threshold extraction in metabolite concentration data}, year = {2004}, abstract = {Motivation: Continued development of analytical techniques based on gas chromatography and mass spectrometry now facilitates the generation of larger sets of metabolite concentration data. An important step towards the understanding of metabolite dynamics is the recognition of stable states where metabolite concentrations exhibit a simple behaviour. Such states can be characterized through the identification of significant thresholds in the concentrations. But general techniques for finding discretization thresholds in continuous data prove to be practically insufficient for detecting states due to the weak conditional dependences in concentration data. Results: We introduce a method of recognizing states in the framework of decision tree induction. It is based upon a global analysis of decision forests where stability and quality are evaluated. It leads to the detection of thresholds that are both comprehensible and robust. Applied to metabolite concentration data, this method has led to the discovery of hidden states in the corresponding variables. Some of these reflect known properties of the biological experiments, and others point to putative new states}, language = {en} } @article{FloeterNicolasSchaubetal.2003, author = {Fl{\"o}ter, Andr{\´e} and Nicolas, Jacques and Schaub, Torsten H. and Selbig, Joachim}, title = {Threshold extraction in metabolite concentration data}, year = {2003}, language = {en} } @phdthesis{Apfelbache2009, author = {Apfelbache, R{\´e}my}, title = {Tolerierbare Inkonsistenzen in Konzeptbeschreibungen}, address = {Potsdam}, pages = {iv, 182 S. : graph. Darst.}, year = {2009}, language = {de} }