@article{KrasnovaVeltriSpengleretal.2013, author = {Krasnova, Hanna and Veltri, Natasha F. and Spengler, Klaus and G{\"u}nther, Oliver}, title = {"Deal of the Day" Platforms what drives Consumer loyalty?}, series = {Business \& information systems engineering : the international journal of Wirtschaftsinformatik}, volume = {5}, journal = {Business \& information systems engineering : the international journal of Wirtschaftsinformatik}, number = {3}, publisher = {Springer}, address = {Heidelberg}, issn = {1867-0202}, doi = {10.1007/s12599-013-0268-2}, pages = {165 -- 177}, year = {2013}, abstract = {"Deal of the Day" (DoD) platforms have quickly become popular by offering savings on local services, products and vacations. For merchants, these platforms represent a new marketing channel to advertise their products and services and attract new customers. DoD platform providers, however, struggle to maintaining a stable market share and profitability, because entry and switching costs are low. To sustain a competitive market position, DoD providers are looking for ways to build a loyal customer base. However, research examining the determinants of user loyalty in this novel context is scarce. To fill this gap, this study employs Grounded Theory methodology to develop a conceptual model of customer loyalty to a DoD provider. In the next step, qualitative insights are enriched and validated using quantitative data from a survey of 202 DoD users. The authors find that customer loyalty is in large part driven by monetary incentives, but can be eroded if impressions from merchant encounters are below expectations. In addition, enhancing the share of deals relevant for consumers, i.e. signal-to-noise ratio, and mitigating perceived risks of a transaction emerge as challenges. Beyond theoretical value, the results offer practical insights into how customer loyalty to a DoD provider can be promoted.}, language = {en} } @article{BrainGebserPuehreretal.2007, author = {Brain, Martin and Gebser, Martin and P{\"u}hrer, J{\"o}rg and Schaub, Torsten H. and Tompits, Hans and Woltran, Stefan}, title = {"That is illogical, Captain!" : the debugging support tool spock for answer-set programs ; system description}, year = {2007}, language = {en} } @article{JuergensenKonstantinidis2006, author = {J{\"u}rgensen, Helmut and Konstantinidis, Stavros}, title = {(Near-)inverses of sequences}, issn = {0020-7160}, doi = {10.1080/00207160500537801}, year = {2006}, abstract = {We introduce the notion of a near-inverse of a non-decreasing sequence of positive integers; near-inverses are intended to assume the role of inverses in cases when the latter cannot exist. We prove that the near-inverse of such a sequence is unique; moreover, the relation of being near-inverses of each other is symmetric, i.e. if sequence g is the near-inverse of sequence f, then f is the near-inverse of g. There is a connection, by approximations, between near- inverses of sequences and inverses of continuous strictly increasing real-valued functions which can be exploited to derive simple expressions for near-inverses}, language = {en} } @article{DelgrandeSchaubTompitsetal.2004, author = {Delgrande, James Patrick and Schaub, Torsten H. and Tompits, Hans and Wang, Kewen}, title = {A classification and survey of preference handling approchaches in nonmonotonic reasoning}, issn = {0824-7935}, year = {2004}, abstract = {In recent years, there has been a large amount of disparate work concerning the representation and reasoning with qualitative preferential information by means of approaches to nonmonotonic reasoning. Given the variety of underlying systems, assumptions, motivations, and intuitions, it is difficult to compare or relate one approach with another. Here, we present an overview and classification for approaches to dealing with preference. A set of criteria for classifying approaches is given, followed by a set of desiderata that an approach might be expected to satisfy. A comprehensive set of approaches is subsequently given and classified with respect to these sets of underlying principles}, language = {en} } @article{Wang2001, author = {Wang, Kewen}, title = {A comparative study of disjunctive well-founded semantics}, isbn = {3-540-42593-4}, year = {2001}, language = {en} } @article{SchaubWang2001, author = {Schaub, Torsten H. and Wang, Kewen}, title = {A comparative study of logic programs with preference}, year = {2001}, language = {en} } @article{SchaubWang2001, author = {Schaub, Torsten H. and Wang, Kewen}, title = {A comparative study of logic programs with preference}, isbn = {1-558-60777-3}, issn = {1045-0823}, year = {2001}, language = {en} } @article{DelgrandeSchaubTompits2000, author = {Delgrande, James Patrick and Schaub, Torsten H. and Tompits, Hans}, title = {A compilation of Brewka and Eiter's approach to prioritizationtion}, isbn = {3-540-41131-3}, year = {2000}, language = {en} } @article{SarsakovSchaubTompitsetal.2004, author = {Sarsakov, Vladimir and Schaub, Torsten H. and Tompits, Hans and Woltran, Stefan}, title = {A compiler for nested logic programming}, isbn = {3-540- 20721-x}, year = {2004}, language = {en} } @article{DelgrandeSchaubTompits2000, author = {Delgrande, James Patrick and Schaub, Torsten H. and Tompits, Hans}, title = {A compiler for ordered logic programs}, year = {2000}, language = {en} } @article{DelgrandeSchaub2003, author = {Delgrande, James Patrick and Schaub, Torsten H.}, title = {A concictency-based paradigm for belief change}, issn = {0004-3702}, year = {2003}, language = {en} } @article{BrueningSchaub2000, author = {Br{\"u}ning, Stefan and Schaub, Torsten H.}, title = {A connection calculus for handling incomplete information}, year = {2000}, language = {en} } @article{DelgrandeSchaub2007, author = {Delgrande, James Patrick and Schaub, Torsten H.}, title = {A consistency-based framework for merging knowledge bases}, issn = {1570-8683}, year = {2007}, language = {en} } @article{DelgrandeSchaub2000, author = {Delgrande, James Patrick and Schaub, Torsten H.}, title = {A consistency-based model for belief change: preliminary report}, isbn = {0-262-51112-6}, year = {2000}, language = {en} } @article{DelgrandeSchaub2000, author = {Delgrande, James Patrick and Schaub, Torsten H.}, title = {A consistency-based model for belief change: preliminary report}, year = {2000}, language = {en} } @article{BesnardSchaub1993, author = {Besnard, Philippe and Schaub, Torsten H.}, title = {A context-based framework for default logics}, isbn = {0-262-51071-5}, year = {1993}, language = {en} } @article{AndjelkovicSimevskiChenetal.2022, author = {Andjelkovic, Marko and Simevski, Aleksandar and Chen, Junchao and Schrape, Oliver and Stamenkovic, Zoran and Krstić, Miloš and Ilic, Stefan and Ristic, Goran and Jaksic, Aleksandar and Vasovic, Nikola and Duane, Russell and Palma, Alberto J. and Lallena, Antonio M. and Carvajal, Miguel A.}, title = {A design concept for radiation hardened RADFET readout system for space applications}, series = {Microprocessors and microsystems}, volume = {90}, journal = {Microprocessors and microsystems}, publisher = {Elsevier}, address = {Amsterdam}, issn = {0141-9331}, doi = {10.1016/j.micpro.2022.104486}, pages = {18}, year = {2022}, abstract = {Instruments for measuring the absorbed dose and dose rate under radiation exposure, known as radiation dosimeters, are indispensable in space missions. They are composed of radiation sensors that generate current or voltage response when exposed to ionizing radiation, and processing electronics for computing the absorbed dose and dose rate. Among a wide range of existing radiation sensors, the Radiation Sensitive Field Effect Transistors (RADFETs) have unique advantages for absorbed dose measurement, and a proven record of successful exploitation in space missions. It has been shown that the RADFETs may be also used for the dose rate monitoring. In that regard, we propose a unique design concept that supports the simultaneous operation of a single RADFET as absorbed dose and dose rate monitor. This enables to reduce the cost of implementation, since the need for other types of radiation sensors can be minimized or eliminated. For processing the RADFET's response we propose a readout system composed of analog signal conditioner (ASC) and a self-adaptive multiprocessing system-on-chip (MPSoC). The soft error rate of MPSoC is monitored in real time with embedded sensors, allowing the autonomous switching between three operating modes (high-performance, de-stress and fault-tolerant), according to the application requirements and radiation conditions.}, language = {en} } @article{DelgrandeSchaubTompits2003, author = {Delgrande, James Patrick and Schaub, Torsten H. and Tompits, Hans}, title = {A framework for compiling preferences in logic programs}, year = {2003}, language = {en} } @article{KreowskyStabernack2021, author = {Kreowsky, Philipp and Stabernack, Christian Benno}, title = {A full-featured FPGA-based pipelined architecture for SIFT extraction}, series = {IEEE access : practical research, open solutions / Institute of Electrical and Electronics Engineers}, volume = {9}, journal = {IEEE access : practical research, open solutions / Institute of Electrical and Electronics Engineers}, publisher = {Inst. of Electr. and Electronics Engineers}, address = {New York, NY}, issn = {2169-3536}, doi = {10.1109/ACCESS.2021.3104387}, pages = {128564 -- 128573}, year = {2021}, abstract = {Image feature detection is a key task in computer vision. Scale Invariant Feature Transform (SIFT) is a prevalent and well known algorithm for robust feature detection. However, it is computationally demanding and software implementations are not applicable for real-time performance. In this paper, a versatile and pipelined hardware implementation is proposed, that is capable of computing keypoints and rotation invariant descriptors on-chip. All computations are performed in single precision floating-point format which makes it possible to implement the original algorithm with little alteration. Various rotation resolutions and filter kernel sizes are supported for images of any resolution up to ultra-high definition. For full high definition images, 84 fps can be processed. Ultra high definition images can be processed at 21 fps.}, language = {en} } @article{DelgrandeSchaubTompits2007, author = {Delgrande, James Patrick and Schaub, Torsten H. and Tompits, Hans}, title = {A general framework for expressing preferences in causal reasoning and planning}, issn = {0955-792X}, doi = {10.1093/logcom/exm046}, year = {2007}, abstract = {We consider the problem of representing arbitrary preferences in causal reasoning and planning systems. In planning, a preference may be seen as a goal or constraint that is desirable, but not necessary, to satisfy. To begin, we define a very general query language for histories, or interleaved sequences of world states and actions. Based on this, we specify a second language in which preferences are defined. A single preference defines a binary relation on histories, indicating that one history is preferred to the other. From this, one can define global preference orderings on the set of histories, the maximal elements of which are the preferred histories. The approach is very general and flexible; thus it constitutes a base language in terms of which higher-level preferences may be defined. To this end, we investigate two fundamental types of preferences that we call choice and temporal preferences. We consider concrete strategies for these types of preferences and encode them in terms of our framework. We suggest how to express aggregates in the approach, allowing, e.g. the expression of a preference for histories with lowest total action costs. Last, our approach can be used to express other approaches and so serves as a common framework in which such approaches can be expressed and compared. We illustrate this by indicating how an approach due to Son and Pontelli can be encoded in our approach, as well as the language PDDL3.}, language = {en} } @article{DelgrandeSchaubTompits2001, author = {Delgrande, James Patrick and Schaub, Torsten H. and Tompits, Hans}, title = {A generic compiler for ordered logic programs}, isbn = {3-540-42593-4}, year = {2001}, language = {en} } @article{AngerKonczakLinkeetal.2005, author = {Anger, Christian and Konczak, Kathrin and Linke, Thomas and Schaub, Torsten H.}, title = {A Glimpse of Answer Set Programming}, issn = {0170-4516}, year = {2005}, language = {en} } @article{TiwariPrakashGrossetal.2020, author = {Tiwari, Abhishek and Prakash, Jyoti and Groß, Sascha and Hammer, Christian}, title = {A large scale analysis of Android}, series = {The journal of systems and software}, volume = {170}, journal = {The journal of systems and software}, publisher = {Elsevier}, address = {New York}, issn = {0164-1212}, doi = {10.1016/j.jss.2020.110775}, pages = {17}, year = {2020}, abstract = {Many Android applications embed webpages via WebView components and execute JavaScript code within Android. Hybrid applications leverage dedicated APIs to load a resource and render it in a WebView. Furthermore, Android objects can be shared with the JavaScript world. However, bridging the interfaces of the Android and JavaScript world might also incur severe security threats: Potentially untrusted webpages and their JavaScript might interfere with the Android environment and its access to native features. No general analysis is currently available to assess the implications of such hybrid apps bridging the two worlds. To understand the semantics and effects of hybrid apps, we perform a large-scale study on the usage of the hybridization APIs in the wild. We analyze and categorize the parameters to hybridization APIs for 7,500 randomly selected and the 196 most popular applications from the Google Playstore as well as 1000 malware samples. Our results advance the general understanding of hybrid applications, as well as implications for potential program analyses, and the current security situation: We discovered thousands of flows of sensitive data from Android to JavaScript, the vast majority of which could flow to potentially untrustworthy code. Our analysis identified numerous web pages embedding vulnerabilities, which we exemplarily exploited. Additionally, we discovered a multitude of applications in which potentially untrusted JavaScript code may interfere with (trusted) Android objects, both in benign and malign applications.}, language = {en} } @article{HlawiczkaGoesselSogomonyan1997, author = {Hlawiczka, A. and G{\"o}ssel, Michael and Sogomonyan, Egor S.}, title = {A linear code-preserving signature analyzer COPMISR}, isbn = {0-8186-7810-0}, year = {1997}, language = {en} } @article{Arnold2007, author = {Arnold, Holger}, title = {A linearized DPLL calculus with learning}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-15421}, year = {2007}, abstract = {This paper describes the proof calculus LD for clausal propositional logic, which is a linearized form of the well-known DPLL calculus extended by clause learning. It is motivated by the demand to model how current SAT solvers built on clause learning are working, while abstracting from decision heuristics and implementation details. The calculus is proved sound and terminating. Further, it is shown that both the original DPLL calculus and the conflict-directed backtracking calculus with clause learning, as it is implemented in many current SAT solvers, are complete and proof-confluent instances of the LD calculus.}, language = {en} } @article{BesnardHunter2001, author = {Besnard, Philippe and Hunter, Anthony}, title = {A logic-based theory of deductive arguments}, issn = {0004-3702}, year = {2001}, language = {en} } @article{SaposhnikovSaposhnikovGoesseletal.1999, author = {Saposhnikov, V. V. and Saposhnikov, Vl. V. and G{\"o}ssel, Michael and Morosov, Andrej}, title = {A method of construction of combinational self-checking units with detection of all single faults}, year = {1999}, language = {en} } @article{BrueningSchaub1996, author = {Br{\"u}ning, Stefan and Schaub, Torsten H.}, title = {A model-based approach to consistency-checking}, isbn = {3-540-61286-6}, year = {1996}, language = {en} } @article{DelgrandeSchaubTompitsetal.2013, author = {Delgrande, James and Schaub, Torsten H. and Tompits, Hans and Woltran, Stefan}, title = {A model-theoretic approach to belief change in answer set programming}, series = {ACM transactions on computational logic}, volume = {14}, journal = {ACM transactions on computational logic}, number = {2}, publisher = {Association for Computing Machinery}, address = {New York}, issn = {1529-3785}, doi = {10.1145/2480759.2480766}, pages = {46}, year = {2013}, abstract = {We address the problem of belief change in (nonmonotonic) logic programming under answer set semantics. Our formal techniques are analogous to those of distance-based belief revision in propositional logic. In particular, we build upon the model theory of logic programs furnished by SE interpretations, where an SE interpretation is a model of a logic program in the same way that a classical interpretation is a model of a propositional formula. Hence we extend techniques from the area of belief revision based on distance between models to belief change in logic programs. We first consider belief revision: for logic programs P and Q, the goal is to determine a program R that corresponds to the revision of P by Q, denoted P * Q. We investigate several operators, including (logic program) expansion and two revision operators based on the distance between the SE models of logic programs. It proves to be the case that expansion is an interesting operator in its own right, unlike in classical belief revision where it is relatively uninteresting. Expansion and revision are shown to satisfy a suite of interesting properties; in particular, our revision operators satisfy all or nearly all of the AGM postulates for revision. We next consider approaches for merging a set of logic programs, P-1,...,P-n. Again, our formal techniques are based on notions of relative distance between the SE models of the logic programs. Two approaches are examined. The first informally selects for each program P-i those models of P-i that vary the least from models of the other programs. The second approach informally selects those models of a program P-0 that are closest to the models of programs P-1,...,P-n. In this case, P-0 can be thought of as a set of database integrity constraints. We examine these operators with regards to how they satisfy relevant postulate sets. Last, we present encodings for computing the revision as well as the merging of logic programs within the same logic programming framework. This gives rise to a direct implementation of our approach in terms of off-the-shelf answer set solvers. These encodings also reflect the fact that our change operators do not increase the complexity of the base formalism.}, language = {en} } @article{SogomonyanSinghGoessel1999, author = {Sogomonyan, Egor S. and Singh, Adit D. and G{\"o}ssel, Michael}, title = {A multi-mode scannable memory element for high test application efficiency and delay testing}, year = {1999}, language = {en} } @article{SogomonyanSinghGoessel1998, author = {Sogomonyan, Egor S. and Singh, Adit D. and G{\"o}ssel, Michael}, title = {A multi-mode scannable memory element for high test application efficiency and delay testing}, year = {1998}, language = {en} } @article{Gohlke1995, author = {Gohlke, Mario}, title = {A new approach for model-based recognition using colour regions}, year = {1995}, language = {en} } @article{SaposhnikovMorosovSaposhnikovetal.1998, author = {Saposhnikov, V. V. and Morosov, Andrej and Saposhnikov, Vl. V. and G{\"o}ssel, Michael}, title = {A new design method for self-checking unidirectional combinational circuits}, year = {1998}, language = {en} } @article{Richter1995, author = {Richter, Peter}, title = {A new deterministic approach for the optimization of cable layouts for power supply systems}, year = {1995}, language = {en} } @article{Goessel1999, author = {G{\"o}ssel, Michael}, title = {A new method of redundancy addition for circuit optimization}, series = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, volume = {1999, 08}, journal = {Preprint / Universit{\"a}t Potsdam, Institut f{\"u}r Informatik}, publisher = {Univ.}, address = {Potsdam}, issn = {0946-7580}, pages = {9 Bl.}, year = {1999}, language = {en} } @article{SogomonyanGoessel1995, author = {Sogomonyan, Egor S. and G{\"o}ssel, Michael}, title = {A new parity preserving multi-input signature analyser}, year = {1995}, language = {en} } @article{GoesselSogomonyan1996, author = {G{\"o}ssel, Michael and Sogomonyan, Egor S.}, title = {A new self-testing parity checker for ultra-reliable applications}, year = {1996}, language = {en} } @article{GoesselSogomonyanMorosov1999, author = {G{\"o}ssel, Michael and Sogomonyan, Egor S. and Morosov, Andrej}, title = {A new totally error propagating compactor for arbitrary cores with digital interfaces}, year = {1999}, language = {en} } @article{KawanabeBlanchardSugiyamaetal.2006, author = {Kawanabe, Motoaki and Blanchard, Gilles and Sugiyama, Masashi and Spokoiny, Vladimir G. and M{\"u}ller, Klaus-Robert}, title = {A novel dimension reduction procedure for searching non-Gaussian subspaces}, issn = {0302-9743}, doi = {10.1007/11679363_19}, year = {2006}, abstract = {In this article, we consider high-dimensional data which contains a low-dimensional non-Gaussian structure contaminated with Gaussian noise and propose a new linear method to identify the non-Gaussian subspace. Our method NGCA (Non-Gaussian Component Analysis) is based on a very general semi-parametric framework and has a theoretical guarantee that the estimation error of finding the non-Gaussian components tends to zero at a parametric rate. NGCA can be used not only as preprocessing for ICA, but also for extracting and visualizing more general structures like clusters. A numerical study demonstrates the usefulness of our method}, language = {en} } @article{LiangLiuLiuetal.2015, author = {Liang, Feng and Liu, Yunzhen and Liu, Hai and Ma, Shilong and Schnor, Bettina}, title = {A Parallel Job Execution Time Estimation Approach Based on User Submission Patterns within Computational Grids}, series = {International journal of parallel programming}, volume = {43}, journal = {International journal of parallel programming}, number = {3}, publisher = {Springer}, address = {New York}, issn = {0885-7458}, doi = {10.1007/s10766-013-0294-1}, pages = {440 -- 454}, year = {2015}, abstract = {Scheduling performance in computational grid can potentially benefit a lot from accurate execution time estimation for parallel jobs. Most existing approaches for the parallel job execution time estimation, however, require ample past job traces and the explicit correlations between the job execution time and the outer layout parameters such as the consumed processor numbers, the user-estimated execution time and the job ID, which are hard to obtain or reveal. This paper presents and evaluates a novel execution time estimation approach for parallel jobs, the user-behavior clustering for execution time estimation, which can give more accurate execution time estimation for parallel jobs through exploring the job similarity and revealing the user submission patterns. Experiment results show that compared to the state-of-art algorithms, our approach can improve the accuracy of the job execution time estimation up to 5.6 \%, meanwhile the time that our approach spends on calculation can be reduced up to 3.8 \%.}, language = {en} } @article{GoesselSogomonyan1996, author = {G{\"o}ssel, Michael and Sogomonyan, Egor S.}, title = {A parity-preserving multi-input signature analyzer and it application for concurrent checking and BIST}, year = {1996}, language = {en} } @article{PearceSarsakovSchaubetal.2002, author = {Pearce, David and Sarsakov, Vladimir and Schaub, Torsten H. and Tompits, Hans and Woltran, Stefan}, title = {A polynomial translation of logic programs with nested expressions into disjunctive logic programs}, isbn = {3-540-43930-7}, year = {2002}, language = {en} } @article{PearceSarsakovSchaubetal.2002, author = {Pearce, David and Sarsakov, Vladimir and Schaub, Torsten H. and Tompits, Hans and Woltran, Stefan}, title = {A polynomial translation of logic programs with nested expressions into disjunctive logic programs : preliminary report}, year = {2002}, language = {en} } @article{DelgrandeSchaubTompits2007, author = {Delgrande, James Patrick and Schaub, Torsten H. and Tompits, Hans}, title = {A preference-based framework for updating logic programs}, isbn = {978-3-540- 72199-4}, year = {2007}, language = {en} } @article{DelgrandeSchaubTompits2006, author = {Delgrande, James Patrick and Schaub, Torsten H. and Tompits, Hans}, title = {A Preference-Based Framework for Updating logic Programs : preliminary reports}, year = {2006}, language = {en} } @article{Boerner1998, author = {B{\"o}rner, Ferdinand}, title = {A remark on the finite lattice representation problem}, year = {1998}, language = {en} } @article{Schoebel2010, author = {Sch{\"o}bel, Michael}, title = {A runtime environment for online processing of operating system kernel events}, isbn = {978-3-86956-036-6}, year = {2010}, language = {en} } @article{SogomonyanSinghGoessel1998, author = {Sogomonyan, Egor S. and Singh, Adit D. and G{\"o}ssel, Michael}, title = {A scan based concrrent BIST approach for low cost on-line testing}, year = {1998}, language = {en} } @article{SchaubWang2003, author = {Schaub, Torsten H. and Wang, Kewen}, title = {A semantic framework for prefernce handling in answer set programming}, year = {2003}, language = {en} } @article{GoesselChakrabartyOcheretnijetal.2004, author = {Goessel, Michael and Chakrabarty, Krishnendu and Ocheretnij, V. and Leininger, Andreas}, title = {A signature analysis technique for the identification of failing vectors with application to Scan-BIST}, issn = {0923-8174}, year = {2004}, abstract = {We present a new technique for uniquely identifying a single failing vector in an interval of test vectors. This technique is applicable to combinational circuits and for scan-BIST in sequential circuits with multiple scan chains. The proposed method relies on the linearity properties of the MISR and on the use of two test sequences, which are both applied to the circuit under test. The second test sequence is derived from the first in a straightforward manner and the same test pattern source is used for both test sequences. If an interval contains only a single failing vector, the algebraic analysis is guaranteed to identify it. We also show analytically that if an interval contains two failing vectors, the probability that this case is interpreted as one failing vector is very low. We present experimental results for the ISCAS benchmark circuits to demonstrate the use of the proposed method for identifying failing test vectors}, language = {en} }