@article{Noack2014, author = {Noack, Franziska}, title = {CREADED: Colored-Relief application for digital elevation data}, series = {Process design for natural scientists: an agile model-driven approach}, journal = {Process design for natural scientists: an agile model-driven approach}, number = {500}, publisher = {Springer}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {186 -- 199}, year = {2014}, abstract = {In the geoinformatics field, remote sensing data is often used for analyzing the characteristics of the current investigation area. This includes DEMs, which are simple raster grids containing grey scales representing the respective elevation values. The project CREADED that is presented in this paper aims at making these monochrome raster images more significant and more intuitively interpretable. For this purpose, an executable interactive model for creating a colored and relief-shaded Digital Elevation Model (DEM) has been designed using the jABC framework. The process is based on standard jABC-SIBs and SIBs that provide specific GIS functions, which are available as Web services, command line tools and scripts.}, language = {en} } @article{Respondek2014, author = {Respondek, Tobias}, title = {A workflow for computing potential areas for wind turbines}, series = {Process design for natural scientists: an agile model-driven approach}, journal = {Process design for natural scientists: an agile model-driven approach}, number = {500}, publisher = {Springer}, address = {Berlin}, isbn = {978-3-662-45005-5}, pages = {200 -- 215}, year = {2014}, abstract = {This paper describes the implementation of a workflow model for service-oriented computing of potential areas for wind turbines in jABC. By implementing a re-executable model the manual effort of a multi-criteria site analysis can be reduced. The aim is to determine the shift of typical geoprocessing tools of geographic information systems (GIS) from the desktop to the web. The analysis is based on a vector data set and mainly uses web services of the "Center for Spatial Information Science and Systems" (CSISS). This paper discusses effort, benefits and problems associated with the use of the web services.}, language = {en} } @article{Scheele2014, author = {Scheele, Lasse}, title = {Location analysis for placing artificial reefs}, series = {Process design for natural scientists: an agile model-driven approach}, journal = {Process design for natural scientists: an agile model-driven approach}, number = {500}, publisher = {Springer}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {216 -- 228}, year = {2014}, abstract = {Location analyses are among the most common tasks while working with spatial data and geographic information systems. Automating the most frequently used procedures is therefore an important aspect of improving their usability. In this context, this project aims to design and implement a workflow, providing some basic tools for a location analysis. For the implementation with jABC, the workflow was applied to the problem of finding a suitable location for placing an artificial reef. For this analysis three parameters (bathymetry, slope and grain size of the ground material) were taken into account, processed, and visualized with the The Generic Mapping Tools (GMT), which were integrated into the workflow as jETI-SIBs. The implemented workflow thereby showed that the approach to combine jABC with GMT resulted in an user-centric yet user-friendly tool with high-quality cartographic outputs.}, language = {en} } @article{Holler2014, author = {Holler, Robin}, title = {GraffDok - a graffiti documentation application}, series = {Process design for natural scientists: an agile model-driven approach}, journal = {Process design for natural scientists: an agile model-driven approach}, number = {500}, publisher = {Springer}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {239 -- 251}, year = {2014}, abstract = {GraffDok is an application helping to maintain an overview over sprayed images somewhere in a city. At the time of writing it aims at vandalism rather than at beautiful photographic graffiti in an underpass. Looking at hundreds of tags and scribbles on monuments, house walls, etc. it would be interesting to not only record them in writing but even make them accessible electronically, including images. GraffDok's workflow is simple and only requires an EXIF-GPS-tagged photograph of a graffito. It automatically determines its location by using reverse geocoding with the given GPS-coordinates and the Gisgraphy WebService. While asking the user for some more meta data, GraffDok analyses the image in parallel with this and tries to detect fore- and background - before extracting the drawing lines and make them stand alone. The command line based tool ImageMagick is used here as well as for accessing EXIF data. Any meta data is written to csv-files, which will stay easily accessible and can be integrated in TeX-files as well. The latter ones are converted to PDF at the end of the workflow, containing a table about all graffiti and a summary for each - including the generated characteristic graffiti pattern image.}, language = {en} } @article{Reso2014, author = {Reso, Judith}, title = {Protein Classification Workflow}, series = {Process Design for Natural Scientists: an agile model-driven approach}, journal = {Process Design for Natural Scientists: an agile model-driven approach}, number = {500}, editor = {Lambrecht, Anna-Lena and Margaria, Tiziana}, publisher = {Springer Verlag}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {65 -- 72}, year = {2014}, abstract = {The protein classification workflow described in this report enables users to get information about a novel protein sequence automatically. The information is derived by different bioinformatic analysis tools which calculate or predict features of a protein sequence. Also, databases are used to compare the novel sequence with known proteins.}, language = {en} } @article{Schulze2014, author = {Schulze, Gunnar}, title = {Workflow for rapid metagenome analysis}, series = {Process design for natural scientists: an agile model-driven approach}, journal = {Process design for natural scientists: an agile model-driven approach}, number = {500}, publisher = {Springer}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {88 -- 100}, year = {2014}, abstract = {Analyses of metagenomes in life sciences present new opportunities as well as challenges to the scientific community and call for advanced computational methods and workflows. The large amount of data collected from samples via next-generation sequencing (NGS) technologies render manual approaches to sequence comparison and annotation unsuitable. Rather, fast and efficient computational pipelines are needed to provide comprehensive statistics and summaries and enable the researcher to choose appropriate tools for more specific analyses. The workflow presented here builds upon previous pipelines designed for automated clustering and annotation of raw sequence reads obtained from next-generation sequencing technologies such as 454 and Illumina. Employing specialized algorithms, the sequence reads are processed at three different levels. First, raw reads are clustered at high similarity cutoff to yield clusters which can be exported as multifasta files for further analyses. Independently, open reading frames (ORFs) are predicted from raw reads and clustered at two strictness levels to yield sets of non-redundant sequences and ORF families. Furthermore, single ORFs are annotated by performing searches against the Pfam database}, language = {en} } @article{Vierheller2014, author = {Vierheller, Janine}, title = {Exploratory Data Analysis}, series = {Process Design for Natural Scientists: an agile model-driven approach}, journal = {Process Design for Natural Scientists: an agile model-driven approach}, number = {500}, editor = {Lambrecht, Anna-Lena and Margaria, Tiziana}, publisher = {Axel Springer Verlag}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {110 -- 126}, year = {2014}, abstract = {In bioinformatics the term exploratory data analysis refers to different methods to get an overview of large biological data sets. Hence, it helps to create a framework for further analysis and hypothesis testing. The workflow facilitates this first important step of the data analysis created by high-throughput technologies. The results are different plots showing the structure of the measurements. The goal of the workflow is the automatization of the exploratory data analysis, but also the flexibility should be guaranteed. The basic tool is the free software R.}, language = {en} } @article{Schuett2014, author = {Sch{\"u}tt, Christine}, title = {Identification of differentially expressed genes}, series = {Process design for natural scientists: an agile model-driven approach}, journal = {Process design for natural scientists: an agile model-driven approach}, number = {500}, publisher = {Springer}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {127 -- 139}, year = {2014}, abstract = {With the jABC it is possible to realize workflows for numerous questions in different fields. The goal of this project was to create a workflow for the identification of differentially expressed genes. This is of special interest in biology, for it gives the opportunity to get a better insight in cellular changes due to exogenous stress, diseases and so on. With the knowledge that can be derived from the differentially expressed genes in diseased tissues, it becomes possible to find new targets for treatment.}, language = {en} } @article{Kuntzsch2014, author = {Kuntzsch, Christian}, title = {Visualization of data transfer paths}, series = {Process design for natural scientists: an agile model-driven approach}, journal = {Process design for natural scientists: an agile model-driven approach}, number = {500}, publisher = {Springer}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {140 -- 148}, year = {2014}, abstract = {A workflow for visualizing server connections using the Google Maps API was built in the jABC. It makes use of three basic services: An XML-based IP address geolocation web service, a command line tool and the Static Maps API. The result of the workflow is an URL leading to an image file of a map, showing server connections between a client and a target host.}, language = {en} } @article{Hibbe2014, author = {Hibbe, Marcel}, title = {Spotlocator - Guess Where the Photo Was Taken!}, series = {Process Design for Natural Scientists: an agile model-driven approach}, journal = {Process Design for Natural Scientists: an agile model-driven approach}, number = {500}, editor = {Lambrecht, Anna-Lena and Margaria, Tiziana}, publisher = {Springer Verlag}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {149 -- 160}, year = {2014}, abstract = {Spotlocator is a game wherein people have to guess the spots of where photos were taken. The photos of a defined area for each game are from panoramio.com. They are published at http://spotlocator. drupalgardens.com with an ID. Everyone can guess the photo spots by sending a special tweet via Twitter that contains the hashtag \#spotlocator, the guessed coordinates and the ID of the photo. An evaluation is published for all tweets. The players are informed about the distance to the real photo spots and the positions are shown on a map.}, language = {en} } @article{Blaese2014, author = {Blaese, Leif}, title = {Data mining for unidentified protein squences}, series = {Process design for natural scientists: an agile model-driven approach}, journal = {Process design for natural scientists: an agile model-driven approach}, number = {500}, publisher = {Springer}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {73 -- 87}, year = {2014}, abstract = {Through the use of next generation sequencing (NGS) technology, a lot of newly sequenced organisms are now available. Annotating those genes is one of the most challenging tasks in sequence biology. Here, we present an automated workflow to find homologue proteins, annotate sequences according to function and create a three-dimensional model.}, language = {en} } @article{Lis2014, author = {Lis, Monika}, title = {Constructing a Phylogenetic Tree}, series = {Process Design for Natural Scientists: an agile model-driven approach}, journal = {Process Design for Natural Scientists: an agile model-driven approach}, number = {500}, editor = {Lambrecht, Anna-Lena and Margaria, Tiziana}, publisher = {Springer Verlag}, address = {Berlin}, isbn = {978-3-662-45005-5}, issn = {1865-0929}, pages = {101 -- 109}, year = {2014}, abstract = {In this project I constructed a workflow that takes a DNA sequence as input and provides a phylogenetic tree, consisting of the input sequence and other sequences which were found during a database search. In this phylogenetic tree the sequences are arranged depending on similarities. In bioinformatics, constructing phylogenetic trees is often used to explore the evolutionary relationships of genes or organisms and to understand the mechanisms of evolution itself.}, language = {en} } @article{FroitzheimBergnerSchroeder2015, author = {Froitzheim, Manuel and Bergner, Nadine and Schroeder, Ulrik}, title = {Android-Workshop zur Vertiefung der Kenntnisse bez{\"u}glich Datenstrukturen und Programmierung in der Studieneingangsphase}, series = {HDI 2014 : Gestalten von {\"U}berg{\"a}ngen}, volume = {2015}, journal = {HDI 2014 : Gestalten von {\"U}berg{\"a}ngen}, number = {9}, editor = {Schwill, Andreas}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-80247}, pages = {11 -- 26}, year = {2015}, abstract = {Die Studieneingangsphase stellt f{\"u}r Studierende eine Schl{\"u}sselphase des terti{\"a}ren Ausbildungsabschnitts dar. Fachwissenschaftliches Wissen wird praxisfern vermittelt und die Studierenden k{\"o}nnen die Zusammenh{\"a}nge zwischen den Themenfeldern der verschiedenen Vorlesungen nicht erkennen. Zur Verbesserung der Situation wurde ein Workshop entwickelt, der die Verbindung der Programmierung und der Datenstrukturen vertieft. Dabei wird das Spiel Go-Moku1 als Android-App von den Studierenden selbst{\"a}ndig entwickelt. Die Kombination aus Software (Java, Android-SDK) und Hardware (Tablet-Computer) f{\"u}r ein kleines realistisches Softwareprojekt stellt f{\"u}r die Studierenden eine neue Erfahrung dar.}, language = {de} } @article{ShenoyKrauledatBlankertzetal.2006, author = {Shenoy, Pradeep and Krauledat, Matthias and Blankertz, Benjamin and Rao, Rajesh P. N. and M{\"u}ller, Klaus-Robert}, title = {Towards adaptive classification for BCI}, doi = {10.1088/1741-2560/3/1/R02}, year = {2006}, abstract = {Non-stationarities are ubiquitous in EEG signals. They are especially apparent in the use of EEG-based brain- computer interfaces (BCIs): (a) in the differences between the initial calibration measurement and the online operation of a BCI, or (b) caused by changes in the subject's brain processes during an experiment (e.g. due to fatigue, change of task involvement, etc). In this paper, we quantify for the first time such systematic evidence of statistical differences in data recorded during offline and online sessions. Furthermore, we propose novel techniques of investigating and visualizing data distributions, which are particularly useful for the analysis of (non-) stationarities. Our study shows that the brain signals used for control can change substantially from the offline calibration sessions to online control, and also within a single session. In addition to this general characterization of the signals, we propose several adaptive classification schemes and study their performance on data recorded during online experiments. An encouraging result of our study is that surprisingly simple adaptive methods in combination with an offline feature selection scheme can significantly increase BCI performance}, language = {en} } @article{Bobda2009, author = {Bobda, Christophe}, title = {Special issue on ReCoSoC 2007 : editorial}, issn = {0141-9331}, doi = {10.1016/j.micpro.2009.01.001}, year = {2009}, language = {en} } @article{BlankertzDornhegeKrauledatetal.2006, author = {Blankertz, Benjamin and Dornhege, Guido and Krauledat, Matthias and M{\"u}ller, Klaus-Robert and Kunzmann, Volker and Losch, Florian and Curio, Gabriel}, title = {The Berlin brain-computer interface : EEG-based communication without subject training}, issn = {1534-4320}, doi = {10.1109/Tnsre.2006.875557}, year = {2006}, abstract = {The Berlin Brain-Computer Interface (BBCI) project develops a noninvasive BCI system whose key features are 1) the use of well-established motor competences as control paradigms, 2) high-dimensional features from 128-channel electroencephalogram (EEG), and 3) advanced machine learning techniques. As reported earlier, our experiments demonstrate that very high information transfer rates can be achieved using the readiness potential (RP) when predicting the laterality of upcoming left-versus right-hand movements in healthy subjects. A more recent study showed that the RP similarily accompanies phantom movements in arm amputees, but the signal strength decreases with longer loss of the limb. In a complementary approach, oscillatory features are used to discriminate imagined movements (left hand versus right hand versus foot). In a recent feedback study with six healthy subjects with no or very little experience with BCI control, three subjects achieved an information transfer rate above 35 bits per minute (bpm), and further two subjects above 24 and 15 bpm, while one subject could not achieve any BCI control. These results are encouraging for an EEG-based BCI system in untrained subjects that is independent of peripheral nervous system activity and does not rely on evoked potentials even when compared to results with very well-trained subjects operating other BCI systems}, language = {en} } @article{WilligMitschke2006, author = {Willig, Andreas and Mitschke, Robert}, title = {Results of bit error measurements with sensor nodes and casuistic consequences for design of energy-efficient error control schemes}, isbn = {978-3-540-32158-3}, year = {2006}, abstract = {For the proper design of energy-efficient error control schemes some insight into channel error patterns is needed. This paper presents bit error and packet loss measurements taken with sensor nodes running the popular RFM}, language = {en} } @article{RozinatVanderAalst2006, author = {Rozinat, A and Van der Aalst, Wil M. P.}, title = {Conformance testing: Measuring the fit and appropriateness of event logs and process models}, year = {2006}, abstract = {Most information systems log events (e.g., transaction logs, audit traits) to audit and monitor the processes they support. At the same time, many of these processes have been explicitly modeled. For example, SAP R/3 logs events in transaction logs and there are EPCs (Event-driven Process Chains) describing the so-called reference models. These reference models describe how the system should be used. The coexistence of event logs and process models raises an interesting question: "Does the event log conform to the process model and vice versa?". This paper demonstrates that there is not a simple answer to this question. To tackle the problem, we distinguish two dimensions of conformance: fitness (the event log may be the result of the process modeled) and appropriateness (the model is a likely candidate from a structural and behavioral point of view). Different metrics have been defined and a Conformance Checker has been implemented within the ProM Framework}, language = {en} } @article{Konczak2006, author = {Konczak, Kathrin}, title = {Voting Theory in Answer Set Programming}, year = {2006}, language = {en} } @article{GerbserLeeLierler2006, author = {Gerbser, Martin and Lee, Joohyung and Lierler, Yuliya}, title = {Elementary sets for logic programs}, isbn = {978-1-57735-281-5}, year = {2006}, language = {en} } @article{Konczak2006, author = {Konczak, Kathrin}, title = {Weak order equivalence for Logic Programs with Prefernces}, year = {2006}, language = {en} } @article{KonczakVogel2005, author = {Konczak, Kathrin and Vogel, Ralf}, title = {Abduction and Preferences in Linguistics}, year = {2005}, language = {en} } @article{CaludeJurgensen2005, author = {Calude, C. S. and Jurgensen, Helmut}, title = {Is complexity a source of incompleteness?}, issn = {0196-8858}, year = {2005}, abstract = {In this paper we prove Chaitin's "heuristic principle," the theorems of a finitely-specified theory cannot be significantly more complex than the theory itself, for an appropriate measure of complexity. We show that the measure is invariant under the change of the Godel numbering. For this measure, the theorems of a finitely-specified, sound, consistent theory strong enough to formalize arithmetic which is arithmetically sound (like Zermelo-Fraenkel set theory with choice or Peano Arithmetic) have bounded complexity, hence every sentence of the theory which is significantly more complex than the theory is unprovable. Previous results showing that incompleteness is not accidental, but ubiquitous are here reinforced in probabilistic terms: the probability that a true sentence of length n is provable in the theory tends to zero when n tends to infinity, while the probability that a sentence of length n is true is strictly positive. (c) 2004 Elsevier Inc. All rights reserved}, language = {en} } @article{Bordihn2005, author = {Bordihn, Henning}, title = {On the number of components in cooperating distributed grammar systems}, issn = {0304-3975}, year = {2005}, abstract = {It is proved that the number of components in context-free cooperating distributed (CD) grammar systems can be reduced to 3 when they are working in the so-called sf-mode of derivation, which is the cooperation protocol which has been considered first for CD grammar systems. In this derivation mode, a component continues the derivation until and unless there is a nonterminal in the sentential form which cannot be rewritten according to that component. Moreover, it is shown that CD grammar systems in sf-mode with only one component can generate only the context-free languages but they can generate non-context-free languages if two components are used. The sf-mode of derivation is compared with other well-known cooperation protocols with respect to the hierarchies induced by the number of components. (C) 2004 Elsevier B.V. All rights reserved}, language = {en} } @article{BeerenwinkelSingLengaueretal.2005, author = {Beerenwinkel, Niko and Sing, Tobias and Lengauer, Thomas and Rahnenfuhrer, Joerg and Roomp, Kirsten and Savenkov, Igor and Fischer, Roman and Hoffmann, Daniel and Selbig, Joachim and Korn, Klaus and Walter, Hauke and Berg, Thomas and Braun, Patrick and Faetkenheuer, Gerd and Oette, Mark and Rockstroh, Juergen and Kupfer, Bernd and Kaiser, Rolf and Daeumer, Martin}, title = {Computational methods for the design of effective therapies against drug resistant HIV strains}, year = {2005}, abstract = {The development of drug resistance is a major obstacle to successful treatment of HIV infection. The extraordinary replication dynamics of HIV facilitates its escape from selective pressure exerted by the human immune system and by combination drug therapy. We have developed several computational methods whose combined use can support the design of optimal antiretroviral therapies based on viral genomic data}, language = {en} } @article{BrzozowskiJuergensen2005, author = {Brzozowski, J. A. and J{\"u}rgensen, Helmut}, title = {Representation of semiautomata by canonical words and equivalences}, issn = {0129-0541}, year = {2005}, abstract = {We study a novel representation of semiautomata, which is motivated by the method of trace-assertion specifications of software modules. Each state of the semiautomaton is represented by an arbitrary word leading to that state, the canonical word. The transitions of the semiautomaton give rise to a right congruence, the state-equivalence, on the set of input words of the semiautomaton: two words are state-equivalent if and only if they lead to the same state. We present a simple algorithm for finding a set of generators for state-equivalence. Directly from this set of generators, we construct a confluent prefix-rewriting system which permits us to transform any word to its canonical representative. In general, the rewriting system may allow infinite derivations. To address this issue, we impose the condition of prefix-continuity on the set of canonical words. A set is prefix-continuous if, whenever a word w and a prefix u of w axe in the set, then all the prefixes of w longer than u are also in the set. Prefix-continuous sets include prefix-free and prefix-closed sets as special cases. We prove that the rewriting system is Noetherian if and only if the set of canonical words is prefix-continuous. Furthermore, if the set of canonical words is prefix- continuous, then the set of rewriting rules is irredundant. We show that each prefix-continuous canonical set corresponds to a spanning forest of the semiautomaton}, language = {en} } @article{BordihnHolzerKutrib2005, author = {Bordihn, Henning and Holzer, Markus and Kutrib, Martin}, title = {Unsolvability levels of operation problems for subclasses of context-free languages}, issn = {0129-0541}, year = {2005}, abstract = {We investigate the operation problem for linear and deterministic context-free languages: Fix an operation on formal languages. Given linear (deterministic, respectively) context-free languages, is the application of this operation to the given languages still a linear (deterministic, respectively) context-free language? Besides the classical operations, for which the linear and deterministic context-free languages are not closed, we also consider the recently introduced root and power operation. We show non-semidecidability, to be more precise, we show completeness for the second level of the arithmetic hierarchy for all of the aforementioned operations, except for the power operation, if the underlying alphabet contains at least two letters. The result for the power opera, tion solves an open problem stated in Theoret. Comput. Sci. 314 (2004) 445-449}, language = {en} } @article{BruggemeierDovifatKubisch2005, author = {Bruggemeier, M. and Dovifat, A. and Kubisch, D.}, title = {Micropolitical innovation arenas as a tool for analyzing innovation processes in the context of electronic government}, issn = {0937-6429}, year = {2005}, abstract = {E-Government requires technical and organizational innovation. Research has already shown that the respective innovation process is complex and contingent upon specific organizational structures. Managing such innovation processes successfully is difficult. Drawing on assumptions of micropolitical behavior, a framework of innovation arenas is proposed. It supports the analysis of ongoing E-Government projects as well as the ex post investigation of successful or failed projects. Testing this framework in case studies already demonstrates its usefulness for individual actors making strategic choices about change management. Furthermore, the results indicate that many commonly held assumptions about successful change management have to be reconsidered}, language = {en} } @article{MeineckeZieheKurthsetal.2005, author = {Meinecke, Frank C. and Ziehe, Andreas and Kurths, J{\"u}rgen and M{\"u}ller, Klaus-Robert}, title = {Measuring phase synchronization of superimposed signals}, issn = {0031-9007}, year = {2005}, abstract = {Phase synchronization is an important phenomenon that occurs in a wide variety of complex oscillatory processes. Measuring phase synchronization can therefore help to gain fundamental insight into nature. In this Letter we point out that synchronization analysis techniques can detect spurious synchronization, if they are fed with a superposition of signals such as in electroencephalography or magnetoencephalography data. We show how techniques from blind source separation can help to nevertheless measure the true synchronization and avoid such pitfalls}, language = {en} } @article{ScholzKaplanGuyetal.2005, author = {Scholz, Matthias and Kaplan, F. and Guy, C. L. and Kopka, Joachim and Selbig, Joachim}, title = {Non-linear PCA : a missing data approach}, issn = {1367-4803}, year = {2005}, abstract = {Motivation: Visualizing and analysing the potential non-linear structure of a dataset is becoming an important task in molecular biology. This is even more challenging when the data have missing values. Results: Here, we propose an inverse model that performs non-linear principal component analysis (NLPCA) from incomplete datasets. Missing values are ignored while optimizing the model, but can be estimated afterwards. Results are shown for both artificial and experimental datasets. In contrast to linear methods, non-linear methods were able to give better missing value estimations for non-linear structured data. Application: We applied this technique to a time course of metabolite data from a cold stress experiment on the model plant Arabidopsis thaliana, and could approximate the mapping function from any time point to the metabolite responses. Thus, the inverse NLPCA provides greatly improved information for better understanding the complex response to cold stress}, language = {en} } @article{CordesKaiserSelbig2006, author = {Cordes, Frank and Kaiser, Rolf and Selbig, Joachim}, title = {Bioinformatics approach to predicting HIV drug resistance}, issn = {1473-7159}, doi = {10.1586/14737159.6.2.207}, year = {2006}, abstract = {The emergence of drug resistance remains one of the most challenging issues in the treatment of HIV-1 infection. The extreme replication dynamics of HIV facilitates its escape from the selective pressure exerted by the human immune system and by the applied combination drug therapy. This article reviews computational methods whose combined use can support the design of optimal antiretroviral therapies based on viral genotypic and phenotypic data. Genotypic assays are based on the analysis of mutations associated with reduced drug susceptibility, but are difficult to interpret due to the numerous mutations and mutational patterns that confer drug resistance. Phenotypic resistance or susceptibility can be experimentally evaluated by measuring the inhibition of the viral replication in cell culture assays. However, this procedure is expensive and time consuming}, language = {en} } @article{LemmCurioHlushchuketal.2006, author = {Lemm, Steven and Curio, Gabriel and Hlushchuk, Yevhen and M{\"u}ller, Klaus-Robert}, title = {Enhancing the signal-to-noise ratio of ICA-based extracted ERPs}, issn = {0018-9294}, doi = {10.1109/Tbme.2006.870258}, year = {2006}, abstract = {When decomposing single trial electroencephalography it is a challenge to incorporate prior physiological knowledge. Here, we develop a method that uses prior information about the phase-locking property of event-related potentials in a regularization framework to bias a blind source separation algorithm toward an improved separation of single-trial phase-locked responses in terms of an increased signal-to-noise ratio. In particular, we suggest a transformation of the data, using weighted average of the single trial and trial-averaged response, that redirects the focus of source separation methods onto the subspace of event-related potentials. The practical benefit with respect to an improved separation of such components from ongoing background activity and extraneous noise is first illustrated on artificial data and finally verified in a real-world application of extracting single-trial somatosensory evoked potentials from multichannel EEG-recordings}, language = {en} } @article{LaubRothBuhmannetal.2006, author = {Laub, Julian and Roth, Volker and Buhmann, Joachim and M{\"u}ller, Klaus-Robert}, title = {On the information and representation of non-Euclidean pairwise data}, issn = {0031-3203}, doi = {10.1016/j.patcog.2006.04.016}, year = {2006}, abstract = {Two common data representations are mostly used in intelligent data analysis, namely the vectorial and the pairwise representation. Pairwise data which satisfy the restrictive conditions of Euclidean spaces can be faithfully translated into a Euclidean vectorial representation by embedding. Non-metric pairwise data with violations of symmetry, reflexivity or triangle inequality pose a substantial conceptual problem for pattern recognition since the amount of predictive structural information beyond what can be measured by embeddings is unclear. We show by systematic modeling of non-Euclidean pairwise data that there exists metric violations which can carry valuable problem specific information. Furthermore, Euclidean and non-metric data can be unified on the level of structural information contained in the data. Stable component analysis selects linear subspaces which are particularly insensitive to data fluctuations. Experimental results from different domains support our pattern recognition strategy.}, language = {en} } @article{KawanabeBlanchardSugiyamaetal.2006, author = {Kawanabe, Motoaki and Blanchard, Gilles and Sugiyama, Masashi and Spokoiny, Vladimir G. and M{\"u}ller, Klaus-Robert}, title = {A novel dimension reduction procedure for searching non-Gaussian subspaces}, issn = {0302-9743}, doi = {10.1007/11679363_19}, year = {2006}, abstract = {In this article, we consider high-dimensional data which contains a low-dimensional non-Gaussian structure contaminated with Gaussian noise and propose a new linear method to identify the non-Gaussian subspace. Our method NGCA (Non-Gaussian Component Analysis) is based on a very general semi-parametric framework and has a theoretical guarantee that the estimation error of finding the non-Gaussian components tends to zero at a parametric rate. NGCA can be used not only as preprocessing for ICA, but also for extracting and visualizing more general structures like clusters. A numerical study demonstrates the usefulness of our method}, language = {en} } @article{PerniciWeske2006, author = {Pernici, Barbara and Weske, Mathias}, title = {Business process management}, issn = {0169-023X}, year = {2006}, language = {en} } @article{Camales2005, author = {Camales, Renaud}, title = {Explicit formulation of the solution of Hamada-Leray-Wagschal's theorem}, issn = {0034-5318}, year = {2005}, abstract = {In this paper, an explicit formula of the solution of Hainada-Leray-Wagschal's theorem is given. For this, only structure's theorem of finite dimensional determination's function and linear algebra technics developped in [1] are used}, language = {en} } @article{HaegerSchubert2005, author = {H{\"a}ger, Sebastian and Schubert, Wolfgang}, title = {Assoziationen in Softwarearchitekturen}, series = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, volume = {2005, 2}, journal = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, publisher = {Univ.}, address = {Potsdam}, issn = {0946-7580}, pages = {33 S.}, year = {2005}, language = {de} } @article{FaberKonczak2005, author = {Faber, Wolfgang and Konczak, Kathrin}, title = {Strong Equivalence for Logic Programs with Preferences}, year = {2005}, language = {en} } @article{KonczakLang2005, author = {Konczak, Kathrin and Lang, Jerome}, title = {Voting procedures with incomplete preferences}, year = {2005}, language = {en} } @article{KonczakVogel2005, author = {Konczak, Kathrin and Vogel, Ralf}, title = {Abduction and preferences in linguistics : Extended abstract}, issn = {0302-9743}, year = {2005}, language = {en} } @article{GoesselMorozovSapozhnikovetal.2005, author = {Goessel, Michael and Morozov, A. V. and Sapozhnikov, V. V. and Sapozhaikov, Vl. V.}, title = {Checking combinational circuits by the method of logic complement}, issn = {0005-1179}, year = {2005}, abstract = {Design of fully self-testing combinational circuits was considered. A theorem defining the conditions for guaranteed logic complement-based design of fully self-testing circuit was proved. Examples were presented}, language = {en} } @article{MeineckeHarmelingMueller2005, author = {Meinecke, Frank C. and Harmeling, Stefan and M{\"u}ller, Klaus-Robert}, title = {Inlier-based ICA with an application to superimposed images}, issn = {0899-9457}, year = {2005}, abstract = {This paper proposes a new independent component analysis (ICA) method which is able to unmix overcomplete mixtures of sparce or structured signals like speech, music or images. Furthermore, the method is designed to be robust against outliers, which is a favorable feature for ICA algorithms since most of them are extremely sensitive to outliers. Our approach is based on a simple outlier index. However, instead of robustifying an existing algorithm by some outlier rejection technique we show how this index can be used directly to solve the ICA problem for super-Gaussian sources. The resulting inlier-based ICA (IBICA) is outlier-robust by construction and can be used for standard ICA as well as for overcomplete ICA (i.e. more source signals than observed signals). (c) 2005 Wiley Periodicals, Inc}, language = {en} } @article{LemmBlankertzCurioetal.2005, author = {Lemm, Steven and Blankertz, Benjamin and Curio, Gabriel and M{\"u}ller, Klaus-Robert}, title = {Spatio-spectral filters for improving the classification of single trial EEG}, issn = {0018-9294}, year = {2005}, abstract = {Data recorded in electroencephalogram (EEG)-based brain-computer interface experiments is generally very noisy, non-stationary, and contaminated with artifacts that can deteriorate discrimination/classification methods. In this paper, we extend the common spatial pattern (CSP) algorithm with the aim to alleviate these adverse effects. In particular, we suggest an extension of CSP to the state space, which utilizes the method of time delay embedding. As we will show, this allows for individually tuned frequency filters at each electrode position and, thus, yields an improved and more robust machine learning procedure. The advantages of the proposed method over the original CSP method are verified in terms of an improved information transfer rate (bits per trial) on a set of EEG-recordings from experiments of imagined limb movements}, language = {en} } @article{WilligMatheusWolisz2005, author = {Willig, Andreas and Matheus, K. and Wolisz, A.}, title = {Wireless technology in industrial networks}, issn = {0018-9219}, year = {2005}, abstract = {With the success of wireless technologies in consumer electronics, standard wireless technologies are envisioned for the deployment in industrial environments as well. Industrial applications involving mobile subsystems or just the desire to save cabling make wireless technologies attractive. Nevertheless, these applications often have stringent requirements on reliability and timing. In wired environments, timing and reliability are well catered for by fieldbus systems (which are a mature technology designed to enable communication between digital controllers and the sensors and actuators interfacing to a physical process). When wireless links are included, reliability and timing requirements are significantly more difficult to meet, due to the adverse properties of the radio channels. In this paper we thus discuss some key issues coming up in wireless fieldbus and wireless industrial communication systems:1)fundamental problems like achieving timely and reliable transmission despite channel errors; 2) the usage of existing wireless technologies for this specific field of applications; and 3) the creation of hybrid systems in which wireless stations are included into existing wired systems}, language = {en} } @article{WeskevanderAalstVerbeek2004, author = {Weske, Mathias and van der Aalst, Wil M. P. and Verbeek, H. M. W.}, title = {Advances in business process management}, issn = {0169-023X}, year = {2004}, language = {en} } @article{NicolelisBirbaumerMuller2004, author = {Nicolelis, Miguel Angelo L. and Birbaumer, Niels and Muller, K. R.}, title = {Untitled}, issn = {0018-9294}, year = {2004}, language = {de} } @article{Bordihn2004, author = {Bordihn, Henning}, title = {Context-freeness of the power of context-free languages is undecidable}, issn = {0304-3975}, year = {2004}, abstract = {The power of a language L is the set of all powers of the words in L. In this paper, the following decision problem is investigated. Given a context-free language L, is the power of L context-free? We show that this problem is decidable for languages over unary alphabets, but it is undecidable whenever languages over alphabets with at least two letters are considered. (C) 2003 Elsevier B.V. All rights reserved}, language = {en} } @article{DornhegeBlankertzCurioetal.2004, author = {Dornhege, Guido and Blankertz, Benjamin and Curio, Gabriel and M{\"u}ller, Klaus-Robert}, title = {Boosting bit rates in noninvasive EEG single-trial classifications by feature combination and multiclass paradigms}, year = {2004}, abstract = {Noninvasive electroencephalogram (EEG) recordings provide for easy and safe access to human neocortical processes which can be exploited for a brain-computer interface (BCI). At present, however, the use of BCIs is severely limited by low bit-transfer rates. We systematically analyze and develop two recent concepts, both capable of enhancing the information gain from multichannel scalp EEG recordings: 1) the combination of classifiers, each specifically tailored for different physiological phenomena, e.g., slow cortical potential shifts, such as the premovement Bereitschaftspotential or differences in spatio-spectral distributions of brain activity (i.e., focal event-related desynchronizations) and 2) behavioral paradigms inducing the subjects to generate one out of several brain states (multiclass approach) which all bare a distinctive spatio-temporal signature well discriminable in the standard scalp EEG. We derive information-theoretic predictions and demonstrate their relevance in experimental data. We will show that a suitably arranged interaction between these concepts can significantly boost BCI performances}, language = {en} } @article{BlankertzMuellerCurioetal.2004, author = {Blankertz, Benjamin and M{\"u}ller, Klaus-Robert and Curio, Gabriel and Vaughan, Theresa M. and Schalk, Gerwin and Wolpaw, Jonathan R. and Schlogl, Alois and Neuper, Christa and Pfurtscheller, Gert and Hinterberger, Thilo and Schroder, Michael and Birbaumer, Niels}, title = {The BCI competition 2003 : Progress and perspectives in detection and discrimination of EEG single trials}, issn = {0018-9294}, year = {2004}, abstract = {Interest in developing a new method of man-to-machine communication-a brain-computer interface (BCI)-has grown steadily over the past few decades. BCIs create a new communication channel between the brain and an output device by bypassing conventional motor output pathways of nerves and muscles. These systems use signals recorded from the scalp, the surface of the cortex, or from inside the brain to enable users to control a variety of applications including simple word-processing software and orthotics. BCI technology could therefore provide a new communication and control option for individuals who cannot otherwise express their wishes to the outside world. Signal processing and classification methods are essential tools in the development of improved BCI technology. We organized the BCI Competition 2003 to evaluate the current state of the art of these tools. Four laboratories well versed in EEG-based BCI research provided six data sets in a documented format. We made these data sets (i.e., labeled training sets and unlabeled test sets) and their descriptions available on the Internet. The goal in the competition was to maximize the performance measure for the test labels. Researchers worldwide tested their algorithms and competed for the best classification results. This paper describes the six data sets and the results and function of the most successful algorithms}, language = {en} } @article{HarmelingMeineckeMueller2004, author = {Harmeling, Stefan and Meinecke, Frank C. and M{\"u}ller, Klaus-Robert}, title = {Injecting noise for analysing the stability of ICA components}, issn = {0165-1684}, year = {2004}, abstract = {Usually, noise is considered to be destructive. We present a new method that constructively injects noise to assess the reliability and the grouping structure of empirical ICA component estimates. Our method can be viewed as a Monte-Carlo-style approximation of the curvature of some performance measure at the solution. Simulations show that the true root-mean-squared angle distances between the real sources and the source estimates can be approximated well by our method. In a toy experiment, we see that we are also able to reveal the underlying grouping structure of the extracted ICA components. Furthermore, an experiment with fetal ECG data demonstrates that our approach is useful for exploratory data analysis of real-world data. (C) 2003 Elsevier B.V. All rights reserved}, language = {en} } @article{GoesselChakrabartyOcheretnijetal.2004, author = {Goessel, Michael and Chakrabarty, Krishnendu and Ocheretnij, V. and Leininger, Andreas}, title = {A signature analysis technique for the identification of failing vectors with application to Scan-BIST}, issn = {0923-8174}, year = {2004}, abstract = {We present a new technique for uniquely identifying a single failing vector in an interval of test vectors. This technique is applicable to combinational circuits and for scan-BIST in sequential circuits with multiple scan chains. The proposed method relies on the linearity properties of the MISR and on the use of two test sequences, which are both applied to the circuit under test. The second test sequence is derived from the first in a straightforward manner and the same test pattern source is used for both test sequences. If an interval contains only a single failing vector, the algebraic analysis is guaranteed to identify it. We also show analytically that if an interval contains two failing vectors, the probability that this case is interpreted as one failing vector is very low. We present experimental results for the ISCAS benchmark circuits to demonstrate the use of the proposed method for identifying failing test vectors}, language = {en} } @article{MuellerVigarioMeineckeetal.2004, author = {M{\"u}ller, Klaus-Robert and Vigario, R. and Meinecke, Frank C. and Ziehe, Andreas}, title = {Blind source separation techniques for decomposing event-related brain signals}, issn = {0218-1274}, year = {2004}, abstract = {Recently blind source separation (BSS) methods have been highly successful when applied to biomedical data. This paper reviews the concept of BSS and demonstrates its usefulness in the context of event-related MEG measurements. In a first experiment we apply BSS to artifact identification of raw MEG data and discuss how the quality of the resulting independent component projections can be evaluated. The second part of our study considers averaged data of event-related magnetic fields. Here, it is particularly important to monitor and thus avoid possible overfitting due to limited sample size. A stability assessment of the BSS decomposition allows to solve this task and an additional grouping of the BSS components reveals interesting structure, that could ultimately be used for gaining a better physiological modeling of the data}, language = {en} } @article{SugiyamaKawanabeMueller2004, author = {Sugiyama, Masashi and Kawanabe, Motoaki and M{\"u}ller, Klaus-Robert}, title = {Trading variance reduction with unbiasedness : the regularized subspace information criterion for robust model selection in kernel regression}, issn = {0899-7667}, year = {2004}, abstract = {A well-known result by Stein (1956) shows that in particular situations, biased estimators can yield better parameter estimates than their generally preferred unbiased counterparts. This letter follows the same spirit, as we will stabilize the unbiased generalization error estimates by regularization and finally obtain more robust model selection criteria for learning. We trade a small bias against a larger variance reduction, which has the beneficial effect of being more precise on a single training set. We focus on the subspace information criterion (SIC), which is an unbiased estimator of the expected generalization error measured by the reproducing kernel Hilbert space norm. SIC can be applied to the kernel regression, and it was shown in earlier experiments that a small regularization of SIC has a stabilization effect. However, it remained open how to appropriately determine the degree of regularization in SIC. In this article, we derive an unbiased estimator of the expected squared error, between SIC and the expected generalization error and propose determining the degree of regularization of SIC such that the estimator of the expected squared error is minimized. Computer simulations with artificial and real data sets illustrate that the proposed method works effectively for improving the precision of SIC, especially in the high-noise-level cases. We furthermore compare the proposed method to the original SIC, the cross-validation, and an empirical Bayesian method in ridge parameter selection, with good results}, language = {en} } @article{ZieheKawanabeHarmeling2004, author = {Ziehe, Andreas and Kawanabe, Motoaki and Harmeling, Stefan}, title = {Blind separation of post-nonlinear mixtures using linearizing transformations and temporal decorrelation}, issn = {1532-4435}, year = {2004}, abstract = {We propose two methods that reduce the post-nonlinear blind source separation problem (PNL-BSS) to a linear BSS problem. The first method is based on the concept of maximal correlation: we apply the alternating conditional expectation (ACE) algorithm-a powerful technique from nonparametric statistics-to approximately invert the componentwise nonlinear functions. The second method is a Gaussianizing transformation, which is motivated by the fact that linearly mixed signals before nonlinear transformation are approximately Gaussian distributed. This heuristic, but simple and efficient procedure works as good as the ACE method. Using the framework provided by ACE, convergence can be proven. The optimal transformations obtained by ACE coincide with the sought-after inverse functions of the nonlinearitics. After equalizing the nonlinearities, temporal decorrelation separation (TDSEP) allows us to recover the source signals. Numerical simulations testing "ACE-TD" and "Gauss-TD" on realistic examples are performed with excellent results}, language = {en} } @article{Linke2003, author = {Linke, Thomas}, title = {Suitable graphs for answer set programming}, issn = {1613-0073}, year = {2003}, language = {en} } @article{Linke2003, author = {Linke, Thomas}, title = {Using nested logic programs for answer set programming}, issn = {1613-0073}, year = {2003}, language = {en} } @article{AngerKonczakLinke2002, author = {Anger, Christian and Konczak, Kathrin and Linke, Thomas}, title = {NoMoRe: A system for non-monotonic reasoning with logic programs under answer set semantics}, isbn = {3-540-42254-4}, year = {2002}, language = {en} } @article{LinkeAngerKonczak2002, author = {Linke, Thomas and Anger, Christian and Konczak, Kathrin}, title = {More on nomore}, isbn = {3-540-44190-5}, year = {2002}, language = {en} } @article{Arnold2007, author = {Arnold, Holger}, title = {A linearized DPLL calculus with learning}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-15421}, year = {2007}, abstract = {This paper describes the proof calculus LD for clausal propositional logic, which is a linearized form of the well-known DPLL calculus extended by clause learning. It is motivated by the demand to model how current SAT solvers built on clause learning are working, while abstracting from decision heuristics and implementation details. The calculus is proved sound and terminating. Further, it is shown that both the original DPLL calculus and the conflict-directed backtracking calculus with clause learning, as it is implemented in many current SAT solvers, are complete and proof-confluent instances of the LD calculus.}, language = {en} } @article{Frank2012, author = {Frank, Mario}, title = {Axiom relevance decision engine : technical report}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-72128}, year = {2012}, abstract = {This document presents an axiom selection technique for classic first order theorem proving based on the relevance of axioms for the proof of a conjecture. It is based on unifiability of predicates and does not need statistical information like symbol frequency. The scope of the technique is the reduction of the set of axioms and the increase of the amount of provable conjectures in a given time. Since the technique generates a subset of the axiom set, it can be used as a preprocessor for automated theorem proving. This technical report describes the conception, implementation and evaluation of ARDE. The selection method, which is based on a breadth-first graph search by unifiability of predicates, is a weakened form of the connection calculus and uses specialised variants or unifiability to speed up the selection. The implementation of the concept is evaluated with comparison to the results of the world championship of theorem provers of the year 2012 (CASC J6). It is shown that both the theorem prover leanCoP which uses the connection calculus and E which uses equality reasoning, can benefit from the selection approach. Also, the evaluation shows that the concept is applyable for theorem proving problems with thousands of formulae and that the selection is independent from the calculus used by the theorem prover.}, language = {en} } @article{KnobelsdorfKreitz2013, author = {Knobelsdorf, Maria and Kreitz, Christoph}, title = {Ein konstruktivistischer Lehransatz f{\"u}r die Einf{\"u}hrungsveranstaltung der Theoretischen Informatik}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {5}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-64845}, pages = {21 -- 32}, year = {2013}, abstract = {Ausgehend von einem sozial-konstruktivistischen Verst{\"a}ndnis von Lernprozessen und unter der besonderen Ber{\"u}cksichtigung der durch die Bologna-Studienreform angeregten Kompetenzorientierung, haben wir in den letzten Jahren einen hochschuldidaktischen Ansatz f{\"u}r die Einf{\"u}hrungsveranstaltung im Bereich der Theoretischen Informatik an der Universit{\"a}t Potsdam entwickelt und praktisch erprobt. Nach zahlreichen Experimenten und mit einer Durchfallquote von zuletzt 6\% im Wintersemester 2011/2012 haben wir den Eindruck, dass der Ansatz den Studierenden jene Lernumgebung und -anregung bietet, die ihnen hilft, die entsprechenden Fachkompetenzen in der Veranstaltung zu entwickeln. In diesem Artikel stellen wir unseren Ansatz vor und skizzieren abschlie{\"y}end, wie wir diesen im n{\"a}chsten Wintersemester empirisch evaluieren werden.}, language = {de} } @article{Goettel2013, author = {G{\"o}ttel, Timo}, title = {Schnupperveranstaltungen Informatik in der Hochschullandschaft}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {5}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-64860}, pages = {45 -- 55}, year = {2013}, abstract = {Die vorliegende Arbeit er{\"o}rtert die Frage, wie Nachwuchs f{\"u}r das Informatikstudium nachhaltig gesichert werden kann. Dazu werden Befragungen unter Sch{\"u}lerinnen und Sch{\"u}lern (13-16 Jahre), sowie aktuelle Informatik-Schnupperangebote f{\"u}r Sch{\"u}lerinnen und Sch{\"u}ler an deutschsprachigen Hochschulen vorgestellt und untersucht. Diese Gegen{\"u}berstellung zeigt deutlich, dass die Angebote nur bedingt eine breite Zielgruppe ansprechen und dass weitere Formate und Inhalte notwendig sind, um Sch{\"u}lerinnen und Sch{\"u}ler fr{\"u}hzeitig und in voller Breite zu erreichen und f{\"u}r das Informatikstudium zu begeistern. Daraus wird abgeleitet, dass Missverst{\"a}ndnisse und Probleme mit der Informatik im Schulkontext aufgegriffen werden m{\"u}ssen. Das vorgestellte Programm Schulbotschafter Informatik stellt einen m{\"o}glichen Weg dar, um dies zu erreichen und {\"u}bliche Schnupperangebote zu erg{\"a}nzen.}, language = {de} } @article{GoetzBrinda2013, author = {G{\"o}tz, Christian and Brinda, Torsten}, title = {Sind soziale Netzwerke geeignet, um darin f{\"u}r Informatikstudieng{\"a}nge zu werben?}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {5}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-65017}, pages = {137 -- 142}, year = {2013}, abstract = {Durch den bundesweiten R{\"u}ckgang der Sch{\"u}lerzahlen und einer steigenden Zahl von Bildungsangeboten geraten Universit{\"a}ten und Hochschulen in den n{\"a}chsten Jahren weiter in eine Wettbewerbssituation, weshalb sie effektive Marketingmaßnahmen entwickeln m{\"u}ssen, um Sch{\"u}lerinnen und Sch{\"u}ler m{\"o}glichst fr{\"u}hzeitig f{\"u}r das jeweilige Angebot (z. B. Informatik- und informatiknahe Studieng{\"a}nge) zu interessieren. Ein Medium, {\"u}ber das sich potenziell sehr viele Jugendliche erreichen lassen, sind dabei soziale Netzwerke. Diese Arbeit pr{\"a}sentiert Ergebnisse einer Studie unter Informatikstudienanf{\"a}ngerinnen und -anf{\"a}ngern zum Nutzungsverhalten sozialer Netzwerke und zieht Schlussfolgerungen zu deren Eignung als Werbe- und Informationskanal f{\"u}r die Zielgruppe der Informatikinteressierten.}, language = {de} } @article{ApelBergBergneretal.2013, author = {Apel, Rebecca and Berg, Tobias and Bergner, Nadine and Chatti, Mhamed Amine and Holz, Jan and Leicht-Scholten, Carmen and Schroeder, Ulrike}, title = {Ein vierstufiges F{\"o}rderkonzept f{\"u}r die Studieneingangsphase in der Informatik}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {5}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-65025}, pages = {143 -- 148}, year = {2013}, abstract = {Es wird ein vierstufiges F{\"o}rderkonzept f{\"u}r die Studieneingangsphase im Fach Informatik beschrieben, das derzeit im Rahmen des Projekts IGaDtools4MINT an der RWTH Aachen auf der Basis einer Literaturanalyse und eines daraus abgeleiteten Indikatorenkatalogs entwickelt wird.}, language = {de} } @article{SchirmerRick2013, author = {Schirmer, Ingrid and Rick, Detlef}, title = {Pers{\"o}nlichkeitsbildung und informatische Professionalisierung}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {5}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-65053}, pages = {160 -- 169}, year = {2013}, language = {de} } @article{Doellner2005, author = {D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Geospatial digital rights management in geovisualization}, issn = {0008-7041}, year = {2005}, abstract = {Geovisualization offers powerful tools, techniques, and strategies to present, explore, analyze, and manage geoinformation. Interactive geovirtual environments such as virtual 3D maps or virtual 3D city models, however, raise the question how to control geodata usage and distribution. We present a concept for embedding digital rights in geovisualizations. It is based on geo-documents, an object-oriented scheme to specify a wide range of geo visualizations. Geo-documents are assembled by building blocks categorized into presentation, structure, interaction, animation, and Digital Rights Management (DRM) classes. DRM objects allow for defining permissions and constraints for all objects contained in geo-documents. In this way, authors of geo visualizations can control how their geo-documents are used, personalized, and redistributed by users. The strengths of the presented concept include the ability to integrate heterogeneous 2D and 3D geodata within a compact design scheme and the ability to cope with privacy, security, and copyright issues. Embedded digital rights in geovisualizations can be applied to improve the usability of geodata user interfaces, to implement publisher-subscriber communication systems for geodata, and to establish business models for geodata trading systems}, language = {en} } @article{NienhausDoellner2005, author = {Nienhaus, Marc and D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Depicting dynamics using principles of visual art and narration's}, issn = {0272-1716}, year = {2005}, language = {en} } @article{DelikostidisEngelRetsiosetal.2013, author = {Delikostidis, Ioannis and Engel, Juri and Retsios, Bas and van Elzakker, Corne P. J. M. and Kraak, Menno-Jan and D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Increasing the usability of pedestrian navigation interfaces by means of landmark visibility analysis}, series = {The journal of navigation}, volume = {66}, journal = {The journal of navigation}, number = {4}, publisher = {Cambridge Univ. Press}, address = {New York}, issn = {0373-4633}, doi = {10.1017/S0373463313000209}, pages = {523 -- 537}, year = {2013}, abstract = {Communicating location-specific information to pedestrians is a challenging task which can be aided by user-friendly digital technologies. In this paper, landmark visibility analysis, as a means for developing more usable pedestrian navigation systems, is discussed. Using an algorithmic framework for image-based 3D analysis, this method integrates a 3D city model with identified landmarks and produces raster visibility layers for each one. This output enables an Android phone prototype application to indicate the visibility of landmarks from the user's actual position. Tested in the field, the method achieves sufficient accuracy for the context of use and improves navigation efficiency and effectiveness.}, language = {en} } @article{RichterKyprianidisDoellner2013, author = {Richter, Rico and Kyprianidis, Jan Eric and D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Out-of-core GPU-based change detection in massive 3D point clouds}, series = {Transactions in GIS}, volume = {17}, journal = {Transactions in GIS}, number = {5}, publisher = {Wiley-Blackwell}, address = {Hoboken}, issn = {1361-1682}, doi = {10.1111/j.1467-9671.2012.01362.x}, pages = {724 -- 741}, year = {2013}, abstract = {If sites, cities, and landscapes are captured at different points in time using technology such as LiDAR, large collections of 3D point clouds result. Their efficient storage, processing, analysis, and presentation constitute a challenging task because of limited computation, memory, and time resources. In this work, we present an approach to detect changes in massive 3D point clouds based on an out-of-core spatial data structure that is designed to store data acquired at different points in time and to efficiently attribute 3D points with distance information. Based on this data structure, we present and evaluate different processing schemes optimized for performing the calculation on the CPU and GPU. In addition, we present a point-based rendering technique adapted for attributed 3D point clouds, to enable effective out-of-core real-time visualization of the computation results. Our approach enables conclusions to be drawn about temporal changes in large highly accurate 3D geodata sets of a captured area at reasonable preprocessing and rendering times. We evaluate our approach with two data sets from different points in time for the urban area of a city, describe its characteristics, and report on applications.}, language = {en} } @article{ParedesBooAmoretal.2012, author = {Paredes, E. G. and Boo, M. and Amor, M. and Bruguera, J. D. and D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Extended hybrid meshing algorithm for multiresolution terrain models}, series = {International journal of geographical information science}, volume = {26}, journal = {International journal of geographical information science}, number = {5}, publisher = {Routledge, Taylor \& Francis Group}, address = {Abingdon}, issn = {1365-8816}, doi = {10.1080/13658816.2011.615317}, pages = {771 -- 793}, year = {2012}, abstract = {Hybrid terrains are a convenient approach for the representation of digital terrain models, integrating heterogeneous data from different sources. In this article, we present a general, efficient scheme for achieving interactive level-of-detail rendering of hybrid terrain models, without the need for a costly preprocessing or resampling of the original data. The presented method works with hybrid digital terrains combining regular grid data and local high-resolution triangulated irregular networks. Since grid and triangulated irregular network data may belong to different datasets, a straightforward combination of both geometries would lead to meshes with holes and overlapping triangles. Our method generates a single multiresolution model integrating the different parts in a coherent way, by performing an adaptive tessellation of the region between their boundaries. Hence, our solution is one of the few existing approaches for integrating different multiresolution algorithms within the same terrain model, achieving a simple interactive rendering of complex hybrid terrains.}, language = {en} } @article{Hecher2022, author = {Hecher, Markus}, title = {Treewidth-aware reductions of normal ASP to SAT}, series = {Artificial intelligence}, volume = {304}, journal = {Artificial intelligence}, publisher = {Elsevier}, address = {Amsterdam}, issn = {0004-3702}, doi = {10.1016/j.artint.2021.103651}, pages = {24}, year = {2022}, abstract = {Answer Set Programming (ASP) is a paradigm for modeling and solving problems for knowledge representation and reasoning. There are plenty of results dedicated to studying the hardness of (fragments of) ASP. So far, these studies resulted in characterizations in terms of computational complexity as well as in fine-grained insights presented in form of dichotomy-style results, lower bounds when translating to other formalisms like propositional satisfiability (SAT), and even detailed parameterized complexity landscapes. A generic parameter in parameterized complexity originating from graph theory is the socalled treewidth, which in a sense captures structural density of a program. Recently, there was an increase in the number of treewidth-based solvers related to SAT. While there are translations from (normal) ASP to SAT, no reduction that preserves treewidth or at least keeps track of the treewidth increase is known. In this paper we propose a novel reduction from normal ASP to SAT that is aware of the treewidth, and guarantees that a slight increase of treewidth is indeed sufficient. Further, we show a new result establishing that, when considering treewidth, already the fragment of normal ASP is slightly harder than SAT (under reasonable assumptions in computational complexity). This also confirms that our reduction probably cannot be significantly improved and that the slight increase of treewidth is unavoidable. Finally, we present an empirical study of our novel reduction from normal ASP to SAT, where we compare treewidth upper bounds that are obtained via known decomposition heuristics. Overall, our reduction works better with these heuristics than existing translations. (c) 2021 Elsevier B.V. All rights reserved.}, language = {en} } @article{CsuhajVarjuDassowVaszil2010, author = {Csuhaj-Varj{\´u}, Erzs{\´e}bet and Dassow, Juergen and Vaszil, Gy{\"o}rgy}, title = {Variants of competence-based derivations in CD grammar systems}, issn = {0129-0541}, doi = {10.1142/S0129054110007428}, year = {2010}, abstract = {In this paper we introduce and study some new cooperation protocols for cooperating distributed (CD) grammar systems. These derivation modes depend on the number of different nonterminals present in the sentential form obtained when a component finished a derivation phase. This measure describes the competence of the grammar on the string (the competence is high if the number of the different nonterminals is small). It is also a measure of the efficiency of the grammar on the given string (a component is more efficient than another one if it is able to decrease the number of nonterminals in the string to a greater extent). We prove that if the underlying derivation mode is the t-mode derivation, then some variants of these systems determine the class of random context ET0L languages. If these CD grammar systems use the k step limited derivations as underlying derivation mode, then they are able to generate any recursively enumerable language.}, language = {en} } @article{HilseLucke2013, author = {Hilse, Michael and Lucke, Ulrike}, title = {eLiS - E-Learning in Studienbereichen}, series = {E-Learning Symposium 2012 : Aktuelle Anwendungen, innovative Prozesse und neueste Ergebnisse aus der E-Learning-Praxis ; Potsdam, 17. November 2012}, journal = {E-Learning Symposium 2012 : Aktuelle Anwendungen, innovative Prozesse und neueste Ergebnisse aus der E-Learning-Praxis ; Potsdam, 17. November 2012}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, doi = {10.25932/publishup-44227}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-442275}, pages = {73 -- 75}, year = {2013}, abstract = {Deutsche Universit{\"a}ten erweitern ihre E-Learning-Angebote als Service f{\"u}r die Studierenden und Lehrenden. Diese sind je nach Fakult{\"a}t unterschiedlich ausgepr{\"a}gt. Dieser Artikel zeigt, wie durch technische Erweiterung der Infrastruktur, einer Anpassung der Organisationsstruktur und einer gezielten Inhaltsentwicklung eine durchg{\"a}ngige und personalisierbare Lehr- und Lernumgebung (Personal Learning Environment, PLE) geschaffen und damit die Akzeptanz bei den Lehrenden und Studierenden f{\"u}r E-Learning erh{\"o}ht werden kann. Aus der vorausgehenden, systematischen Anforderungsanalyse k{\"o}nnen Kennzahlen f{\"u}r die Qualit{\"a}tssicherung von E-Learning-Angeboten abgeleitet werden.}, language = {de} } @article{ZoernerKoehlmannBrandt2014, author = {Zoerner, Dietmar and K{\"o}hlmann, Wiebke and Brandt, Christopher}, title = {Mobiles spielebasiertes Lernen an historischen Lernorten}, series = {E-Learning Symposium 2014 : Mobil und vernetzt - studieren im digitalen Zeitalter ; Potsdam, 14. November 2014}, journal = {E-Learning Symposium 2014 : Mobil und vernetzt - studieren im digitalen Zeitalter ; Potsdam, 14. November 2014}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, doi = {10.25932/publishup-44235}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-442354}, pages = {53 -- 54}, year = {2014}, abstract = {Im Rahmen eines interdisziplin{\"a}ren studentischen Projekts wurde ein Framework f{\"u}r mobile pervasive Lernspiele entwickelt. Am Beispiel des historischen Lernortes Park Sanssouci wurde auf dieser Grundlage ein Lernspiel f{\"u}r Sch{\"u}lerinnen und Sch{\"u}ler implementiert. Die geplante Evaluation soll die Lernwirksamkeit von geobasierten mobilen Lernspielen messen. Dazu wird die Intensit{\"a}t des Flow-Erlebens mit einer ortsgebundenen alternativen Umsetzung verglichen.}, language = {de} } @article{AlLabanRegerLucke2022, author = {Al Laban, Firas and Reger, Martin and Lucke, Ulrike}, title = {Closing the Policy Gap in the Academic Bridge}, series = {Education sciences}, volume = {12}, journal = {Education sciences}, number = {12}, publisher = {MDPI}, address = {Basel}, issn = {2227-7102}, doi = {10.3390/educsci12120930}, year = {2022}, abstract = {The highly structured nature of the educational sector demands effective policy mechanisms close to the needs of the field. That is why evidence-based policy making, endorsed by the European Commission under Erasmus+ Key Action 3, aims to make an alignment between the domains of policy and practice. Against this background, this article addresses two issues: First, that there is a vertical gap in the translation of higher-level policies to local strategies and regulations. Second, that there is a horizontal gap between educational domains regarding the policy awareness of individual players. This was analyzed in quantitative and qualitative studies with domain experts from the fields of virtual mobility and teacher training. From our findings, we argue that the combination of both gaps puts the academic bridge from secondary to tertiary education at risk, including the associated knowledge proficiency levels. We discuss the role of digitalization in the academic bridge by asking the question: which value does the involved stakeholders expect from educational policies? As a theoretical basis, we rely on the model of value co-creation for and by stakeholders. We describe the used instruments along with the obtained results and proposed benefits. Moreover, we reflect on the methodology applied, and we finally derive recommendations for future academic bridge policies.}, language = {en} } @article{HollmannFrohmeEndrullatetal.2020, author = {Hollmann, Susanne and Frohme, Marcus and Endrullat, Christoph and Kremer, Andreas and D'Elia, Domenica and Regierer, Babette and Nechyporenko, Alina}, title = {Ten simple rules on how to write a standard operating procedure}, series = {PLOS Computational Biology}, volume = {16}, journal = {PLOS Computational Biology}, number = {9}, publisher = {PLOS}, address = {San Francisco}, pages = {10}, year = {2020}, abstract = {Research publications and data nowadays should be publicly available on the internet and, theoretically, usable for everyone to develop further research, products, or services. The long-term accessibility of research data is, therefore, fundamental in the economy of the research production process. However, the availability of data is not sufficient by itself, but also their quality must be verifiable. Measures to ensure reuse and reproducibility need to include the entire research life cycle, from the experimental design to the generation of data, quality control, statistical analysis, interpretation, and validation of the results. Hence, high-quality records, particularly for providing a string of documents for the verifiable origin of data, are essential elements that can act as a certificate for potential users (customers). These records also improve the traceability and transparency of data and processes, therefore, improving the reliability of results. Standards for data acquisition, analysis, and documentation have been fostered in the last decade driven by grassroot initiatives of researchers and organizations such as the Research Data Alliance (RDA). Nevertheless, what is still largely missing in the life science academic research are agreed procedures for complex routine research workflows. Here, well-crafted documentation like standard operating procedures (SOPs) offer clear direction and instructions specifically designed to avoid deviations as an absolute necessity for reproducibility. Therefore, this paper provides a standardized workflow that explains step by step how to write an SOP to be used as a starting point for appropriate research documentation.}, language = {en} } @article{PrescherBornscheinKoehlmannetal.2018, author = {Prescher, Denise and Bornschein, Jens and K{\"o}hlmann, Wiebke and Weber, Gerhard}, title = {Touching graphical applications}, series = {Universal Access in the Information Society}, volume = {17}, journal = {Universal Access in the Information Society}, number = {2}, publisher = {Springer}, address = {Heidelberg}, issn = {1615-5289}, doi = {10.1007/s10209-017-0538-8}, pages = {391 -- 409}, year = {2018}, abstract = {Novel two-dimensional tactile displays enable blind users to not only get access to the textual but also to the graphical content of a graphical user interface. Due to the higher amount of information that can be presented in parallel, orientation and exploration can be more complex. In this paper we present the HyperBraille system, which consists of a pin-matrix device as well as a graphical screen reader providing the user with appropriate presentation and interaction possibilities. To allow for a detailed analysis of bimanual interaction strategies on a pin-matrix device, we conducted two user studies with a total of 12 blind people. The task was to fill in .pdf forms on the pin-matrix device by using different input methods, namely gestures, built-in hardware buttons as well as a conventional PC keyboard. The forms were presented in a semigraphic view type that not only contains Braille but also tactile widgets in a spatial arrangement. While completion time and error rate partly depended on the chosen input method, the usage of special reading strategies seemed to be independent of it. A direct comparison of the system and a conventional assistive technology (screen reader with single-line Braille device) showed that interaction on the pin-matrix device can be very efficient if the user is trained. The two-dimensional output can improve access to .pdf forms with insufficient accessibility as the mapping of input controls and the corresponding labels can be supported by a spatial presentation.}, language = {en} } @article{MetrefCosmeLeSommeretal.2019, author = {Metref, Sammy and Cosme, Emmanuel and Le Sommer, Julien and Poel, Nora and Brankart, Jean-Michel and Verron, Jacques and Gomez Navarro, Laura}, title = {Reduction of spatially structured errors in Wide-Swath altimetric satellite data using data assimilation}, series = {Remote sensing}, volume = {11}, journal = {Remote sensing}, number = {11}, publisher = {MDPI}, address = {Basel}, issn = {2072-4292}, doi = {10.3390/rs11111336}, pages = {21}, year = {2019}, abstract = {The Surface Water and Ocean Topography (SWOT) mission is a next generation satellite mission expected to provide a 2 km-resolution observation of the sea surface height (SSH) on a two-dimensional swath. Processing SWOT data will be challenging because of the large amount of data, the mismatch between a high spatial resolution and a low temporal resolution, and the observation errors. The present paper focuses on the reduction of the spatially structured errors of SWOT SSH data. It investigates a new error reduction method and assesses its performance in an observing system simulation experiment. The proposed error-reduction method first projects the SWOT SSH onto a subspace spanned by the SWOT spatially structured errors. This projection is removed from the SWOT SSH to obtain a detrended SSH. The detrended SSH is then processed within an ensemble data assimilation analysis to retrieve a full SSH field. In the latter step, the detrending is applied to both the SWOT data and an ensemble of model-simulated SSH fields. Numerical experiments are performed with synthetic SWOT observations and an ensemble from a North Atlantic, 1/60 degrees simulation of the ocean circulation (NATL60). The data assimilation analysis is carried out with an ensemble Kalman filter. The results are assessed with root mean square errors, power spectrum density, and spatial coherence. They show that a significant part of the large scale SWOT errors is reduced. The filter analysis also reduces the small scale errors and allows for an accurate recovery of the energy of the signal down to 25 km scales. In addition, using the SWOT nadir data to adjust the SSH detrending further reduces the errors.}, language = {en} } @article{PousttchiGleiss2019, author = {Pousttchi, Key and Gleiß, Alexander}, title = {Surrounded by middlemen - how multi-sided platforms change the insurance industry}, series = {Electron Markets}, volume = {29}, journal = {Electron Markets}, number = {4}, publisher = {Springer}, address = {Heidelberg}, issn = {1019-6781}, doi = {10.1007/s12525-019-00363-w}, pages = {609 -- 629}, year = {2019}, abstract = {Multi-sided platforms (MSP) strongly affect markets and play a crucial part within the digital and networked economy. Although empirical evidence indicates their occurrence in many industries, research has not investigated the game-changing impact of MSP on traditional markets to a sufficient extent. More specifically, we have little knowledge of how MSP affect value creation and customer interaction in entire markets, exploiting the potential of digital technologies to offer new value propositions. Our paper addresses this research gap and provides an initial systematic approach to analyze the impact of MSP on the insurance industry. For this purpose, we analyze the state of the art in research and practice in order to develop a reference model of the value network for the insurance industry. On this basis, we conduct a case-study analysis to discover and analyze roles which are occupied or even newly created by MSP. As a final step, we categorize MSP with regard to their relation to traditional insurance companies, resulting in a classification scheme with four MSP standard types: Competition, Coordination, Cooperation, Collaboration.}, language = {en} } @article{AfantenosPeldszusStede2018, author = {Afantenos, Stergos and Peldszus, Andreas and Stede, Manfred}, title = {Comparing decoding mechanisms for parsing argumentative structures}, series = {Argument \& Computation}, volume = {9}, journal = {Argument \& Computation}, number = {3}, publisher = {IOS Press}, address = {Amsterdam}, issn = {1946-2166}, doi = {10.3233/AAC-180033}, pages = {177 -- 192}, year = {2018}, abstract = {Parsing of argumentative structures has become a very active line of research in recent years. Like discourse parsing or any other natural language task that requires prediction of linguistic structures, most approaches choose to learn a local model and then perform global decoding over the local probability distributions, often imposing constraints that are specific to the task at hand. Specifically for argumentation parsing, two decoding approaches have been recently proposed: Minimum Spanning Trees (MST) and Integer Linear Programming (ILP), following similar trends in discourse parsing. In contrast to discourse parsing though, where trees are not always used as underlying annotation schemes, argumentation structures so far have always been represented with trees. Using the 'argumentative microtext corpus' [in: Argumentation and Reasoned Action: Proceedings of the 1st European Conference on Argumentation, Lisbon 2015 / Vol. 2, College Publications, London, 2016, pp. 801-815] as underlying data and replicating three different decoding mechanisms, in this paper we propose a novel ILP decoder and an extension to our earlier MST work, and then thoroughly compare the approaches. The result is that our new decoder outperforms related work in important respects, and that in general, ILP and MST yield very similar performance.}, language = {en} } @article{BaedkeSchoettler2017, author = {Baedke, Jan and Sch{\"o}ttler, Tobias}, title = {Visual Metaphors in the Sciences}, series = {Journal for General Philosophy of Science}, volume = {48}, journal = {Journal for General Philosophy of Science}, publisher = {Springer}, address = {Dordrecht}, issn = {0925-4560}, doi = {10.1007/s10838-016-9353-9}, pages = {173 -- 194}, year = {2017}, abstract = {Recent philosophical analyses of the epistemic dimension of images in the sciences show a certain trend in acknowledging potential roles of these images beyond their merely decorative or pedagogical functions. We argue, however, that this new debate has yet paid little attention to a special type of pictures, we call 'visual metaphor', and its versatile heuristic potential in organizing data, supporting communication, and guiding research, modeling, and theory formation. Based on a case study of Conrad Hal Waddington's epigenetic landscape images in biology, we develop a descriptive framework applicable to heuristic roles of various visual metaphors in the sciences.}, language = {en} } @article{WaitelonisJuergesSack2019, author = {Waitelonis, J{\"o}rg and J{\"u}rges, Henrik and Sack, Harald}, title = {Remixing entity linking evaluation datasets for focused benchmarking}, series = {Semantic Web}, volume = {10}, journal = {Semantic Web}, number = {2}, publisher = {IOS Press}, address = {Amsterdam}, issn = {1570-0844}, doi = {10.3233/SW-180334}, pages = {385 -- 412}, year = {2019}, abstract = {In recent years, named entity linking (NEL) tools were primarily developed in terms of a general approach, whereas today numerous tools are focusing on specific domains such as e.g. the mapping of persons and organizations only, or the annotation of locations or events in microposts. However, the available benchmark datasets necessary for the evaluation of NEL tools do not reflect this focalizing trend. We have analyzed the evaluation process applied in the NEL benchmarking framework GERBIL [in: Proceedings of the 24th International Conference on World Wide Web (WWW'15), International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, Switzerland, 2015, pp. 1133-1143, Semantic Web 9(5) (2018), 605-625] and all its benchmark datasets. Based on these insights we have extended the GERBIL framework to enable a more fine grained evaluation and in depth analysis of the available benchmark datasets with respect to different emphases. This paper presents the implementation of an adaptive filter for arbitrary entities and customized benchmark creation as well as the automated determination of typical NEL benchmark dataset properties, such as the extent of content-related ambiguity and diversity. These properties are integrated on different levels, which also enables to tailor customized new datasets out of the existing ones by remixing documents based on desired emphases. Besides a new system library to enrich provided NIF [in: International Semantic Web Conference (ISWC'13), Lecture Notes in Computer Science, Vol. 8219, Springer, Berlin, Heidelberg, 2013, pp. 98-113] datasets with statistical information, best practices for dataset remixing are presented, and an in depth analysis of the performance of entity linking systems on special focus datasets is presented.}, language = {en} } @article{GianniniRichterServettoetal.2018, author = {Giannini, Paola and Richter, Tim and Servetto, Marco and Zucca, Elena}, title = {Tracing sharing in an imperative pure calculus}, series = {Science of computer programming}, volume = {172}, journal = {Science of computer programming}, publisher = {Elsevier}, address = {Amsterdam}, issn = {0167-6423}, doi = {10.1016/j.scico.2018.11.007}, pages = {180 -- 202}, year = {2018}, abstract = {We introduce a type and effect system, for an imperative object calculus, which infers sharing possibly introduced by the evaluation of an expression, represented as an equivalence relation among its free variables. This direct representation of sharing effects at the syntactic level allows us to express in a natural way, and to generalize, widely-used notions in literature, notably uniqueness and borrowing. Moreover, the calculus is pure in the sense that reduction is defined on language terms only, since they directly encode store. The advantage of this non-standard execution model with respect to a behaviorally equivalent standard model using a global auxiliary structure is that reachability relations among references are partly encoded by scoping. (C) 2018 Elsevier B.V. All rights reserved.}, language = {en} } @article{Stede2020, author = {Stede, Manfred}, title = {From connectives to coherence relations}, series = {Revue roumaine de linguistique : RRL = Romanian review of linguistics}, volume = {65}, journal = {Revue roumaine de linguistique : RRL = Romanian review of linguistics}, number = {3}, publisher = {Ed. Academiei Rom{\^a}ne}, address = {Bucure{\c{s}}ti}, issn = {0035-3957}, pages = {213 -- 233}, year = {2020}, abstract = {The notion of coherence relations is quite widely accepted in general, but concrete proposals differ considerably on the questions of how they should be motivated, which relations are to be assumed, and how they should be defined. This paper takes a "bottom-up" perspective by assessing the contribution made by linguistic signals (connectives), using insights from the relevant literature as well as verification by practical text annotation. We work primarily with the German language here and focus on the realm of contrast. Thus, we suggest a new inventory of contrastive connective functions and discuss their relationship to contrastive coherence relations that have been proposed in earlier work.}, language = {en} } @article{TiwariPrakashGrossetal.2020, author = {Tiwari, Abhishek and Prakash, Jyoti and Groß, Sascha and Hammer, Christian}, title = {A large scale analysis of Android}, series = {The journal of systems and software}, volume = {170}, journal = {The journal of systems and software}, publisher = {Elsevier}, address = {New York}, issn = {0164-1212}, doi = {10.1016/j.jss.2020.110775}, pages = {17}, year = {2020}, abstract = {Many Android applications embed webpages via WebView components and execute JavaScript code within Android. Hybrid applications leverage dedicated APIs to load a resource and render it in a WebView. Furthermore, Android objects can be shared with the JavaScript world. However, bridging the interfaces of the Android and JavaScript world might also incur severe security threats: Potentially untrusted webpages and their JavaScript might interfere with the Android environment and its access to native features. No general analysis is currently available to assess the implications of such hybrid apps bridging the two worlds. To understand the semantics and effects of hybrid apps, we perform a large-scale study on the usage of the hybridization APIs in the wild. We analyze and categorize the parameters to hybridization APIs for 7,500 randomly selected and the 196 most popular applications from the Google Playstore as well as 1000 malware samples. Our results advance the general understanding of hybrid applications, as well as implications for potential program analyses, and the current security situation: We discovered thousands of flows of sensitive data from Android to JavaScript, the vast majority of which could flow to potentially untrustworthy code. Our analysis identified numerous web pages embedding vulnerabilities, which we exemplarily exploited. Additionally, we discovered a multitude of applications in which potentially untrusted JavaScript code may interfere with (trusted) Android objects, both in benign and malign applications.}, language = {en} } @article{BordihnVaszil2020, author = {Bordihn, Henning and Vaszil, Gy{\"o}rgy}, title = {Deterministic Lindenmayer systems with dynamic control of parallelism}, series = {International journal of foundations of computer science}, volume = {31}, journal = {International journal of foundations of computer science}, number = {1}, publisher = {World Scientific}, address = {Singapore}, issn = {0129-0541}, doi = {10.1142/S0129054120400031}, pages = {37 -- 51}, year = {2020}, abstract = {M-rate 0L systems are interactionless Lindenmayer systems together with a function assigning to every string a set of multisets of productions that may be applied simultaneously to the string. Some questions that have been left open in the forerunner papers are examined, and the computational power of deterministic M-rate 0L systems is investigated, where also tabled and extended variants are taken into consideration.}, language = {en} } @article{ReinhardtMascherGueletal.2010, author = {Reinhardt, Wolfgang and Mascher, Michael and G{\"u}l, Senol and Magenheim, Johannes}, title = {Integration eines Rapid-Feedback-Moduls in eine koaktive Lern- und Arbeitsumgebung}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {4}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-64339}, pages = {53 -- 58}, year = {2010}, abstract = {Die Evaluierung von Lehrveranstaltungen hat in vielen Lehreinrichtungen eine lange Tradition. In diesen klassischen Evaluierungsszenarien werden einmalig pro Semester Umfrageb{\"o}gen an die Studierenden verteilt und anschließend manuell ausgewertet. Die Ergebnisse sind dann zumeist am Ende der Vorlesungszeit vorhanden und geben einen punktuellen Einblick in die Qualit{\"a}t der Lehrveranstaltung bis zum Zeitpunkt der durchgef{\"u}hrten Evaluation. In diesem Artikel stellen wir das Konzept des Rapid Feedback, seine Einsatzm{\"o}glichkeiten in universit{\"a}ren Lehrveranstaltungen und eine prototypische Integration in eine koaktive Lern- und Arbeitsumgebung vor.}, language = {de} } @article{DrewsSchirmerMorisseetal.2013, author = {Drews, Paul and Schirmer, Ingrid and Morisse, Marcel and Sagawe, Arno and Rolf, Arno}, title = {Reflexionsdialog mit DialogueMaps}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {5}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-64983}, pages = {117 -- 123}, year = {2013}, abstract = {In einigen Bereichen der Informatiklehre ist es m{\"o}glich, die pers{\"o}nlichen Erfahrungen der Studierenden im Umgang mit Informationstechnik aufzugreifen und vor dem Hintergrund theoretischer Konzepte aus der Literatur gemeinsam mit ihnen zu reflektieren. Das hier vorgestellte Lehrkonzept des Reflexionsdialogs erstreckt sich {\"u}ber drei Seminartermine und vorbereitende Selbstlernphasen. Unterst{\"u}tzt wird das Konzept durch DialogueMaps, eine Software zur Visualisierung komplexer Sachverhalte und zur Unterst{\"u}tzung interaktiver Dialoge. Dieser Beitrag beschreibt die Hintergr{\"u}nde des Lehrkonzeptes, das Lehrkonzept selbst sowie die inhaltliche Ausgestaltung im Rahmen eines Mastermoduls „Computergest{\"u}tzte Kooperation".}, language = {de} } @article{LaengrichSchulze2015, author = {L{\"a}ngrich, Matthias and Schulze, J{\"o}rg}, title = {Angewandte Output-Orientierung}, series = {HDI 2014 : Gestalten von {\"U}berg{\"a}ngen}, volume = {2015}, journal = {HDI 2014 : Gestalten von {\"U}berg{\"a}ngen}, number = {9}, editor = {Schwill, Andreas and Schubert, Sigrid}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-80299}, pages = {93 -- 107}, year = {2015}, abstract = {Erstsemester-Studierende sind mit den Anforderungen des Lehr-/ Lernprozess einer Universit{\"a}t oder Fachhochschule noch nicht vertraut. Ihre Erwartungen orientieren sich vielmehr an ihrer bisherigen Lerngeschichte (Abitur, Fachabitur, o. {\"a}.). Neben den fachlichen Anforderungen des ersten Semesters m{\"u}ssen die Studierenden also auch Ver{\"a}nderungen im Lehr-/Lernprozess erkennen und bew{\"a}ltigen. Es wird anhand einer Output-orientierten informatischen Lehrveranstaltung aufgezeigt, dass sich aus deren strengen Anforderungen der Messbarkeit klare Kompetenzbeschreibungen ergeben, die besonders dem Orientierungsbed{\"u}rfnis Erstsemester-Studierender entgegenkommen.}, language = {de} } @article{KehrerKelter2010, author = {Kehrer, Timo and Kelter, Udo}, title = {Eine aufwandsbeschr{\"a}nkte Einf{\"u}hrung in die modellbasierte Softwareentwicklung}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {4}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-64304}, pages = {23 -- 33}, year = {2010}, abstract = {Zusammenfassung: Game-based Learning und Edutainment sind aktuelle Schlagworte im Bereich der Hochschulausbildung. Zun{\"a}chst verbindet man damit die Integration einer Spiel- und Spaßkultur in die herk{\"o}mmlichen Lehrveranstaltungen wie Vorlesungen, {\"U}bungen, Praktika und Seminare. Die nachfolgenden Ausf{\"u}hrungen gehen einer genaueren Begriffsanalyse nach und untersuchen, ob Game-based Learning und Edutainment tats{\"a}chlich neuartige Unterrichtsformen erfordern oder neue didaktische {\"U}berlegungen in bestehendes Unterrichtsgeschehen bringen - oder ob es nicht doch an einigen Stellen „alter Wein in neuen Schl{\"a}uchen" ist.}, language = {de} } @article{BordihnVaszil2021, author = {Bordihn, Henning and Vaszil, Gy{\"o}rgy}, title = {Reversible parallel communicating finite automata systems}, series = {Acta informatica}, volume = {58}, journal = {Acta informatica}, number = {4}, publisher = {Springer}, address = {Berlin ; Heidelberg ; New York, NY}, issn = {0001-5903}, doi = {10.1007/s00236-021-00396-9}, pages = {263 -- 279}, year = {2021}, abstract = {We study the concept of reversibility in connection with parallel communicating systems of finite automata (PCFA in short). We define the notion of reversibility in the case of PCFA (also covering the non-deterministic case) and discuss the relationship of the reversibility of the systems and the reversibility of its components. We show that a system can be reversible with non-reversible components, and the other way around, the reversibility of the components does not necessarily imply the reversibility of the system as a whole. We also investigate the computational power of deterministic centralized reversible PCFA. We show that these very simple types of PCFA (returning or non-returning) can recognize regular languages which cannot be accepted by reversible (deterministic) finite automata, and that they can even accept languages that are not context-free. We also separate the deterministic and non-deterministic variants in the case of systems with non-returning communication. We show that there are languages accepted by non-deterministic centralized PCFA, which cannot be recognized by any deterministic variant of the same type.}, language = {en} } @article{BordihnHolzer2021, author = {Bordihn, Henning and Holzer, Markus}, title = {On the number of active states in finite automata}, series = {Acta informatica}, volume = {58}, journal = {Acta informatica}, number = {4}, publisher = {Springer}, address = {Berlin ; Heidelberg [u.a.]}, issn = {0001-5903}, doi = {10.1007/s00236-021-00397-8}, pages = {301 -- 318}, year = {2021}, abstract = {We introduce a new measure of descriptional complexity on finite automata, called the number of active states. Roughly speaking, the number of active states of an automaton A on input w counts the number of different states visited during the most economic computation of the automaton A for the word w. This concept generalizes to finite automata and regular languages in a straightforward way. We show that the number of active states of both finite automata and regular languages is computable, even with respect to nondeterministic finite automata. We further compare the number of active states to related measures for regular languages. In particular, we show incomparability to the radius of regular languages and that the difference between the number of active states and the total number of states needed in finite automata for a regular language can be of exponential order.}, language = {en} } @article{BordihnMitrana2020, author = {Bordihn, Henning and Mitrana, Victor}, title = {On the degrees of non-regularity and non-context-freeness}, series = {Journal of computer and system sciences}, volume = {108}, journal = {Journal of computer and system sciences}, publisher = {Elsevier}, address = {San Diego, Calif. [u.a.]}, issn = {0022-0000}, doi = {10.1016/j.jcss.2019.09.003}, pages = {104 -- 117}, year = {2020}, abstract = {We study the derivational complexity of context-free and context-sensitive grammars by counting the maximal number of non-regular and non-context-free rules used in a derivation, respectively. The degree of non-regularity/non-context-freeness of a language is the minimum degree of non-regularity/non-context-freeness of context-free/context-sensitive grammars generating it. A language has finite degree of non-regularity iff it is regular. We give a condition for deciding whether the degree of non-regularity of a given unambiguous context-free grammar is finite. The problem becomes undecidable for arbitrary linear context-free grammars. The degree of non-regularity of unambiguous context-free grammars generating non-regular languages as well as that of grammars generating deterministic context-free languages that are not regular is of order Omega(n). Context-free non-regular languages of sublinear degree of non-regularity are presented. A language has finite degree of non-context-freeness if it is context-free. Context-sensitive grammars with a quadratic degree of non-context-freeness are more powerful than those of a linear degree.}, language = {en} } @article{GebserJanhunenRintanen2020, author = {Gebser, Martin and Janhunen, Tomi and Rintanen, Jussi}, title = {Declarative encodings of acyclicity properties}, series = {Journal of logic and computation}, volume = {30}, journal = {Journal of logic and computation}, number = {4}, publisher = {Oxford Univ. Press}, address = {Eynsham, Oxford}, issn = {0955-792X}, doi = {10.1093/logcom/exv063}, pages = {923 -- 952}, year = {2020}, abstract = {Many knowledge representation tasks involve trees or similar structures as abstract datatypes. However, devising compact and efficient declarative representations of such structural properties is non-obvious and can be challenging indeed. In this article, we take a number of acyclicity properties into consideration and investigate various logic-based approaches to encode them. We use answer set programming as the primary representation language but also consider mappings to related formalisms, such as propositional logic, difference logic and linear programming. We study the compactness of encodings and the resulting computational performance on benchmarks involving acyclic or tree structures.}, language = {en} } @article{GebserMarateaRicca2020, author = {Gebser, Martin and Maratea, Marco and Ricca, Francesco}, title = {The Seventh Answer Set Programming Competition}, series = {Theory and practice of logic programming}, volume = {20}, journal = {Theory and practice of logic programming}, number = {2}, publisher = {Cambridge Univ. Press}, address = {Cambridge [u.a.]}, issn = {1471-0684}, doi = {10.1017/S1471068419000061}, pages = {176 -- 204}, year = {2020}, abstract = {Answer Set Programming (ASP) is a prominent knowledge representation language with roots in logic programming and non-monotonic reasoning. Biennial ASP competitions are organized in order to furnish challenging benchmark collections and assess the advancement of the state of the art in ASP solving. In this paper, we report on the design and results of the Seventh ASP Competition, jointly organized by the University of Calabria (Italy), the University of Genova (Italy), and the University of Potsdam (Germany), in affiliation with the 14th International Conference on Logic Programming and Non-Monotonic Reasoning (LPNMR 2017).}, language = {en} } @article{GebserMarateaRicca2017, author = {Gebser, Martin and Maratea, Marco and Ricca, Francesco}, title = {The sixth answer set programming competition}, series = {Journal of artificial intelligence research : JAIR}, volume = {60}, journal = {Journal of artificial intelligence research : JAIR}, publisher = {AI Access Found.}, address = {Marina del Rey}, issn = {1076-9757}, doi = {10.1613/jair.5373}, pages = {41 -- 95}, year = {2017}, abstract = {Answer Set Programming (ASP) is a well-known paradigm of declarative programming with roots in logic programming and non-monotonic reasoning. Similar to other closely related problemsolving technologies, such as SAT/SMT, QBF, Planning and Scheduling, advancements in ASP solving are assessed in competition events. In this paper, we report about the design and results of the Sixth ASP Competition, which was jointly organized by the University of Calabria (Italy), Aalto University (Finland), and the University of Genoa (Italy), in affiliation with the 13th International Conference on Logic Programming and Non-Monotonic Reasoning. This edition maintained some of the design decisions introduced in 2014, e.g., the conception of sub-tracks, the scoring scheme,and the adherence to a fixed modeling language in order to push the adoption of the ASP-Core-2 standard. On the other hand, it featured also some novelties, like a benchmark selection stage classifying instances according to their empirical hardness, and a "Marathon" track where the topperforming systems are given more time for solving hard benchmarks.}, language = {en} } @article{KreowskyStabernack2021, author = {Kreowsky, Philipp and Stabernack, Christian Benno}, title = {A full-featured FPGA-based pipelined architecture for SIFT extraction}, series = {IEEE access : practical research, open solutions / Institute of Electrical and Electronics Engineers}, volume = {9}, journal = {IEEE access : practical research, open solutions / Institute of Electrical and Electronics Engineers}, publisher = {Inst. of Electr. and Electronics Engineers}, address = {New York, NY}, issn = {2169-3536}, doi = {10.1109/ACCESS.2021.3104387}, pages = {128564 -- 128573}, year = {2021}, abstract = {Image feature detection is a key task in computer vision. Scale Invariant Feature Transform (SIFT) is a prevalent and well known algorithm for robust feature detection. However, it is computationally demanding and software implementations are not applicable for real-time performance. In this paper, a versatile and pipelined hardware implementation is proposed, that is capable of computing keypoints and rotation invariant descriptors on-chip. All computations are performed in single precision floating-point format which makes it possible to implement the original algorithm with little alteration. Various rotation resolutions and filter kernel sizes are supported for images of any resolution up to ultra-high definition. For full high definition images, 84 fps can be processed. Ultra high definition images can be processed at 21 fps.}, language = {en} } @article{LisecSteinfathMeyeretal.2009, author = {Lisec, Jan and Steinfath, Matthias and Meyer, Rhonda C. and Selbig, Joachim and Melchinger, Albrecht E. and Willmitzer, Lothar and Altmann, Thomas}, title = {Identification of heterotic metabolite QTL in Arabidopsis thaliana RIL and IL populations}, issn = {0960-7412}, doi = {10.1111/j.1365-313X.2009.03910.x}, year = {2009}, abstract = {Two mapping populations of a cross between the Arabidopsis thaliana accessions Col-0 and C24 were cultivated and analyzed with respect to the levels of 181 metabolites to elucidate the biological phenomenon of heterosis at the metabolic level. The relative mid-parent heterosis in the F-1 hybrids was <20\% for most metabolic traits. The first mapping population consisting of 369 recombinant inbred lines (RILs) and their test cross progeny with both parents allowed us to determine the position and effect of 147 quantitative trait loci (QTL) for metabolite absolute mid-parent heterosis (aMPH). Furthermore, we identified 153 and 83 QTL for augmented additive (Z(1)) and dominance effects (Z(2)), respectively. We identified putative candidate genes for these QTL using the ARACYC database (http://www.arabidopsis.org/ biocyc), and calculated the average degree of dominance, which was within the dominance and over-dominance range for most metabolites. Analyzing a second population of 41 introgression lines (ILs) and their test crosses with the recurrent parent, we identified 634 significant differences in metabolite levels. Nine per cent of these effects were classified as over-dominant, according to the mode of inheritance. A comparison of both approaches suggested epistasis as a major contributor to metabolite heterosis in Arabidopsis. A linear combination of metabolite levels was shown to significantly correlate with biomass heterosis (r = 0.62).}, language = {en} } @article{Schulte2009, author = {Schulte, Carsten}, title = {Biographisches Lernen in der Informatik}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {1}, publisher = {Universit{\"a}tsverlag Potsdam}, address = {Potsdam}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-29649}, pages = {47 -- 63}, year = {2009}, abstract = {Biographisches Lernen betont insbesondere die Rolle individueller biographischer Erfahrungen und deren Auswirkungen auf Selbstbild, Weltbild und Verhaltensmuster. Schlagwortartig kann diese Perspektive als Unterschied zwischen ‚Informatik lernen' und ‚Informatiker/in werden' beschrieben werden. Im Artikel wird die Perspektive des Biographischen Lernens an Beispielen aus der Informatik skizziert. Biographisches Lernen ist in der Informatik zun{\"a}chst aus rein pragmatischen Gr{\"u}nden bedeutsam. Der rasche Wandel der Informationstechnologien im Alltag ver{\"a}ndert Erfahrungshintergr{\"u}nde der Studierenden (bzw. Sch{\"u}lerinnen und Sch{\"u}ler). Dementsprechend ver{\"a}ndern sich Erwartungen, Interessen, Vorkenntnisse, generelle Einstellungen oder auch ganz banal die ‚IT-Ausstattung' der Lernenden.}, language = {de} } @article{Keil2009, author = {Keil, Reinhard}, title = {Medi@Thing}, series = {Commentarii informaticae didacticae : (CID)}, journal = {Commentarii informaticae didacticae : (CID)}, number = {1}, issn = {1868-0844}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-29611}, pages = {9 -- 24}, year = {2009}, abstract = {In vielen Bildungsbereichen verschieben sich gegenw{\"a}rtig die Akzente von einer qualifikationsorientierten zu einer st{\"a}rker kompetenzorientierten Ausbildung. Der Begriff der Kompetenz zielt dabei darauf ab, das Aneignen von Wissen auf Vorrat zugunsten des Erwerbs allgemeiner Kompetenzen beispielsweise zur eigenst{\"a}ndigen Aneignung von Wissen zu verschieben. Schl{\"u}sselqualifikationen geraten dabei zunehmend auch f{\"u}r die Ausbildung von Ingenieuren und Informatikern ins Blickfeld. Der kooperative und erw{\"a}gende Umgang mit Wissensvielfalt wird in einer Informationsgesellschaft mit ihrer schnellen und verteilten Aufbereitung von Wissen zur Herausforderung. Der Beitrag skizziert die damit verbundenen Anforderungen und stellt mit dem Konzept des Medi@Thing einen Ansatz vor, der die Umsetzung solcher Anforderungen im Rahmen der universit{\"a}ren Informatikausbildung erm{\"o}glicht. Ein zentrales Moment dieses Ansatzes ist die ko-aktive Wissensarbeit in virtuellen R{\"a}umen. Erste Erfahrungen zeigen, dass das Konzept von Studierenden angenommen wird. Es wird jedoch auch deutlich, dass f{\"u}r das Gelingen bestimmte organisatorische und technische Randbedingungen zu beachten sind.}, language = {de} }