Refine
Has Fulltext
- yes (160)
Document Type
- Conference Proceeding (160) (remove)
Language
- English (160) (remove)
Is part of the Bibliography
- no (160) (remove)
Keywords
- Information Structure (3)
- middleware (3)
- additive particle (2)
- Aspektorientierte Softwareentwicklung (1)
- Betriebssysteme (1)
- Coccinelle (1)
- Constraint Solving (1)
- Contrast (1)
- Correction (1)
- Deduction (1)
Institute
- Extern (118)
- Institut für Künste und Medien (18)
- Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung (10)
- Institut für Physik und Astronomie (9)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (7)
- Institut für Geowissenschaften (6)
- Sonderforschungsbereich 632 - Informationsstruktur (6)
- Institut für Informatik und Computational Science (3)
- Institut für Biochemie und Biologie (2)
- Department Linguistik (1)
In this talk, I would like to share my experiences gained from participating in four CSP solver competitions and the second ASP solver competition. In particular, I’ll talk about how various programming techniques can make huge differences in solving some of the benchmark problems used in the competitions. These techniques include global constraints, table constraints, and problem-specific propagators and labeling strategies for selecting variables and values. I’ll present these techniques with experimental results from B-Prolog and other CLP(FD) systems.
A constraint programming system combines two essential components: a constraint solver and a search engine. The constraint solver reasons about satisfiability of conjunctions of constraints, and the search engine controls the search for solutions by iteratively exploring a disjunctive search tree defined by the constraint program. The Monadic Constraint Programming framework gives a monadic definition of constraint programming where the solver is defined as a monad threaded through the monadic search tree. Search and search strategies can then be defined as firstclass objects that can themselves be built or extended by composable search transformers. Search transformers give a powerful and unifying approach to viewing search in constraint programming, and the resulting constraint programming system is first class and extremely flexible.
Preface
(2010)
The workshops on (constraint) logic programming (WLP) are the annual meeting of the Society of Logic Programming (GLP e.V.) and bring together researchers interested in logic programming, constraint programming, and related areas like databases, artificial intelligence and operations research. In this decade, previous workshops took place in Dresden (2008), Würzburg (2007), Vienna (2006), Ulm (2005), Potsdam (2004), Dresden (2002), Kiel (2001), and Würzburg (2000). Contributions to workshops deal with all theoretical, experimental, and application aspects of constraint programming (CP) and logic programming (LP), including foundations of constraint/ logic programming. Some of the special topics are constraint solving and optimization, extensions of functional logic programming, deductive databases, data mining, nonmonotonic reasoning, , interaction of CP/LP with other formalisms like agents, XML, JAVA, program analysis, program transformation, program verification, meta programming, parallelism and concurrency, answer set programming, implementation and software techniques (e.g., types, modularity, design patterns), applications (e.g., in production, environment, education, internet), constraint/logic programming for semantic web systems and applications, reasoning on the semantic web, data modelling for the web, semistructured data, and web query languages.
An important characteristic of Service-Oriented Architectures is that clients do not depend on the service implementation's internal assignment of methods to objects. It is perhaps the most important technical characteristic that differentiates them from more common object-oriented solutions. This characteristic makes clients and services malleable, allowing them to be rearranged at run-time as circumstances change. That improvement in malleability is impaired by requiring clients to direct service requests to particular services. Ideally, the clients are totally oblivious to the service structure, as they are to aspect structure in aspect-oriented software. Removing knowledge of a method implementation's location, whether in object or service, requires re-defining the boundary line between programming language and middleware, making clearer specification of dependence on protocols, and bringing the transaction-like concept of failure scopes into language semantics as well. This paper explores consequences and advantages of a transition from object-request brokering to service-request brokering, including the potential to improve our ability to write more parallel software.
Aspect-oriented middleware is a promising technology for the realisation of dynamic reconfiguration in heterogeneous distributed systems. However, like other dynamic reconfiguration approaches, AO-middleware-based reconfiguration requires that the consistency of the system is maintained across reconfigurations. AO-middleware-based reconfiguration is an ongoing research topic and several consistency approaches have been proposed. However, most of these approaches tend to be targeted at specific contexts, whereas for distributed systems it is crucial to cover a wide range of operating conditions. In this paper we propose an approach that offers distributed, dynamic reconfiguration in a consistent manner, and features a flexible framework-based consistency management approach to cover a wide range of operating conditions. We evaluate our approach by investigating the configurability and transparency of our approach and also quantify the performance overheads of the associated consistency mechanisms.
Enforcing security policies to distributed systems is difficult, in particular, when a system contains untrusted components. We designed AspectKE*, a distributed AOP language based on a tuple space, to tackle this issue. In AspectKE*, aspects can enforce access control policies that depend on future behavior of running processes. One of the key language features is the predicates and functions that extract results of static program analysis, which are useful for defining security aspects that have to know about future behavior of a program. AspectKE* also provides a novel variable binding mechanism for pointcuts, so that pointcuts can uniformly specify join points based on both static and dynamic information about the program. Our implementation strategy performs fundamental static analysis at load-time, so as to retain runtime overheads minimal. We implemented a compiler for AspectKE*, and demonstrate usefulness of AspectKE* through a security aspect for a distributed chat system.
Component based software development (CBSD) and aspectoriented software development (AOSD) are two complementary approaches. However, existing proposals for integrating aspects into component models are direct transposition of object-oriented AOSD techniques to components. In this article, we propose a new approach based on views. Our proposal introduces crosscutting components quite naturally and can be integrated into different component models.
Because software development is increasingly expensive and timeconsuming, software reuse gains importance. Aspect-oriented software development modularizes crosscutting concerns which enables their systematic reuse. Literature provides a number of AOP patterns and best practices for developing reusable aspects based on compelling examples for concerns like tracing, transactions and persistence. However, such best practices are lacking for systematically reusing invasive aspects. In this paper, we present the ‘callback mismatch problem’. This problem arises in the context of abstraction mismatch, in which the aspect is required to issue a callback to the base application. As a consequence, the composition of invasive aspects is cumbersome to implement, difficult to maintain and impossible to reuse. We motivate this problem in a real-world example, show that it persists in the current state-of-the-art, and outline the need for advanced aspectual composition mechanisms to deal with this.
Preface
(2010)
Aspect-oriented programming, component models, and design patterns are modern and actively evolving techniques for improving the modularization of complex software. In particular, these techniques hold great promise for the development of "systems infrastructure" software, e.g., application servers, middleware, virtual machines, compilers, operating systems, and other software that provides general services for higher-level applications. The developers of infrastructure software are faced with increasing demands from application programmers needing higher-level support for application development. Meeting these demands requires careful use of software modularization techniques, since infrastructural concerns are notoriously hard to modularize. Aspects, components, and patterns provide very different means to deal with infrastructure software, but despite their differences, they have much in common. For instance, component models try to free the developer from the need to deal directly with services like security or transactions. These are primary examples of crosscutting concerns, and modularizing such concerns are the main target of aspect-oriented languages. Similarly, design patterns like Visitor and Interceptor facilitate the clean modularization of otherwise tangled concerns. Building on the ACP4IS meetings at AOSD 2002-2009, this workshop aims to provide a highly interactive forum for researchers and developers to discuss the application of and relationships between aspects, components, and patterns within modern infrastructure software. The goal is to put aspects, components, and patterns into a common reference frame and to build connections between the software engineering and systems communities.
Large open-source software projects involve developers with a wide variety of backgrounds and expertise. Such software projects furthermore include many internal APIs that developers must understand and use properly. According to the intended purpose of these APIs, they are more or less frequently used, and used by developers with more or less expertise. In this paper, we study the impact of usage patterns and developer expertise on the rate of defects occurring in the use of internal APIs. For this preliminary study, we focus on memory management APIs in the Linux kernel, as the use of these has been shown to be highly error prone in previous work. We study defect rates and developer expertise, to consider e.g., whether widely used APIs are more defect prone because they are used by less experienced developers, or whether defects in widely used APIs are more likely to be fixed.
The paper aims to bring the experience of playing videogames closer to objective knowledge, where the experience can be assessed and falsified via an operational concept. The theory focuses on explaining the basic elements that form the core of the process of the experience. The name of puppetry is introduced after discussing the similarities in the importance of experience for both videogames and theatrical puppetry. Puppetry, then, operationalizes the gaming experience into a concept that can be assessed.
This paper explores the role of the intentional stance in games, arguing that any question of artificial intelligence has as much to do with the co-option of the player’s interpretation of actions as intelligent as any actual fixed-state systems attached to agents. It demonstrates how simply using a few simple and, in system terms, cheap tricks, existing AI can be both supported and enhanced. This includes representational characteristics, importing behavioral expectations from real life, constraining these expectations using diegetic devices, and managing social interrelationships to create the illusion of a greater intelligence than is ever actually present. It is concluded that complex artificial intelligence is often of less importance to the experience of intelligent agents in play than the creation of a space where the intentional stance can be evoked and supported.
This paper approaches the debate over the notion of “magic circle” through an exploratory analysis of the unfolding of identities/differences in gameplay through Derrida’s différance. Initially, différance is related to the notion of play and identity/difference in Derrida’s perspective. Next, the notion of magic circle through Derrida’s play is analyzed, emphasizing the dynamics of différance to understand gameplay as process; questioning its boundaries. Finally, the focus shifts toward the implications of the interplay of identities and differences during gameplay.
Being "in the game"
(2008)
When people describe themselves as being “in the game” this is often thought to mean they have a sense of presence, i.e. they feel like they are in the virtual environment (Brown/Cairns 2004). Presence research traditionally focuses on user experiences in virtual reality systems (e.g. head mounted displays, CAVE-like systems). In contrast, the experience of gaming is very different. Gamers willingly submit to the rules of the game, learn arbitrary relationships between the controls and the screen output, and take on the persona of their game character. Also whereas presence in VR systems is immediate, presence in gaming is gradual. Due to these differences, one can question the extent to which people feel present during gaming. A qualitative study was conducted to explore what gamers actually mean when they describe themselves as being “in the game.” Thirteen gamers were interviewed and the resulting grounded theory suggests being “in the game” does not necessarily mean presence (i.e. feeling like you are the character and present in the VE). Some people use this phrase just to emphasize their high involvement in the game. These findings differ with Brown and Cairns as they suggest at the highest state of immersion not everybody experiences presence. Furthermore, the experience of presence does not appear dependent on the game being in the first person perspective or the gamer being able to empathize with the character. Future research should investigate why some people experience presence and others do not. Possible explanations include: use of language, perception of presence, personality traits, and types of immersion.
MMORPGs such as WORLD OF WARCRAFT can be understood as interactive representations of war. Within the frame provided by the program the players experience martial conflicts and thus a “virtual war.” The game world however requires a technical and as far as possible invisible infrastructure which has to be protected against attacks: Infrastructure means e.g. the servers on which the data of the player characters and the game’s world are saved, as well as the user accounts, which have to be protected, among other things, from “identity theft.” Besides the war on the virtual surface of the program we will therefore describe the invisible war concerning the infrastructure, the outbreak of which is always feared by the developers and operators of online-worlds, requiring them to take precautions. Furthermore we would like to focus on “virtual game worlds” as places of complete surveillance. Since action in these worlds is always associated with the production of data, total observation is theoretically possible and put into practice by the so-called “game master.” The observation of different communication channels (including user forums) serves to monitor and direct the actions on the virtual battlefield subtly, without the player feeling that his freedom is being limited. Finally, we will compare the fictional theater of war in WORLD OF WARCRAFT to the vision of “Network-Centric Warfare,” since it has often been observed that the analysis of MMORPGs is useful to the real trade of war. However, we point out what an unrealistic theater of war WORLD OF WARCRAFT really is.
Playing with information : how political games encourage the player to cross the magic circle
(2008)
The concept of the magic circle suggests that the experience of play is separated from reality. However, in order to interact with a game’s rule system, the player has to make meaningful interpretations of its representations – and representations are never neutral. Games with political content refer in their representations explicitly to social discourses. Cues within their representational layers provoke the player to link the experience of play to mental concepts of reality.
This paper focuses on the way computer games refer to the context of their formation and ask how they might stimulate the user’s understanding of the world around him. The central question is: Do computer games have the potential to inspire our reflection about moral and ethical issues? And if so, by which means do they achieve this? Drawing on concepts of the ethical criticism in literary studies as proposed by Wayne C. Booth and Martha Nussbaum, I will argue in favor of an ethical criticism for computer games. Two aspects will be brought into focus: the ethical reflection in the artifact as a whole, and the recipient’s emotional involvement. The paper aims at evaluating the interaction of game content and game structure in order to give an adequate insight into the way computer games function and affect us.
Metacommunicative circles
(2008)
The paper uses Gregory Bateson’s concept of metacommunication to explore the boundaries of the ‘magic circle’ in play and computer games. It argues that the idea of a self-contained “magic circle” ignores the constant negotiations among players which establish the realm of play. The “magic circle” is no fixed ontological entity but is set up by metacommunicative play. The paper further pursues the question if metacommunication could also be found in single-player computer games, and comes to the conclusion that metacommunication is implemented in single-player games by the means of metalepsis.
Extending Alexander Galloway’s analysis of the action-image in videogames, this essay explores the concept in relation to its source: the analysis of cinema by the French philosopher Gilles Deleuze. The applicability of the concept to videogames will, therefore, be considered through a comparison between the First Person Shooter S.T.A.L.K.E.R. and Andrey Tarkovsky’s film Stalker. This analysis will compellingly explore the nature of videogame-action, its relation to player-perceptions and its location within the machinic and ludic schema.
This paper suggests an approach to studying the rhetoric of persuasive computer games through comparative analysis. A comparison of the military propaganda game AMERICA’S ARMY to similar shooter games reveals an emphasis on discipline and constraints in all main aspects of the games, demonstrating a preoccupation with ethos more than pathos. Generalizing from this, a model for understanding game rhetoric through balances of freedom and constraints is proposed.
Most play spaces support completely different actions than we normally would think of when moving through real space, out of play. This paper therefore discusses the relationship between selected game rules and game spaces in connection to the behaviors, or possible behaviors, of the player. Space will be seen as a modifier or catalyst of player behavior. Six categories of game space are covered: Joy of movement, exploration, tactical, social, performative, and creative spaces. Joy of movement is examined in detail, with a briefer explanation of the other categories.
This text compares the special characteristics of the game space in computer-generated environments with that in non-computerized playing-situations. Herewith, the concept of the magic circle as a deliberately delineated playing sphere with specific rules to be upheld by the players, is challenged. Yet, computer games also provide a virtual playing environment containing the rules of the game as well as the various action possibilities. But both the hardware and software facilitate the player’s actions rather than constraining them. This makes computer games fundamentally different: in contrast to traditional game spaces or limits, the computer-generated environment does not rely on the awareness of the player in upholding these rules. – Thus, there is no magic circle.
Landscape aesthetics drawing on philosophy and psychology allow us to understand computer games from a new angle. The landscapes of computer games can be understood as environments or images. This difference creates two options: 1. We experience environments or images, or 2. We experience landscape simultaneously as both. Psychologically, the first option can be backed up by a Vygotskian framework (this option highlights certain non-mainstream subject positions), the second by a Piegatian (highlighting cognitive mapping of game worlds).
This paper highlights the different ways of perceiving video games and video game content, incorporating interactive and non-interactive methods. It examines varying cognitive and emotive reactions by persons who are used to play video games as well as persons who are unfamiliar with the aesthetics and the most basic game play rules incorporated within video games. Additionally, the principle of “Flow” serves as a theoretical and philosophical foundation. A small case-study featuring two games has been made to emphasize the numerous possible ways of perception of video games.
In a common description, to play a game is to step inside a concrete or metaphorical magic circle where special rules apply. In video game studies, this description has received an inordinate amount of criticism which the paper argues has two primary sources: 1. a misreading of the basic concept of the magic circle and 2. a somewhat rushed application of traditional theoretical concerns onto games. The paper argues that games studies must move beyond conventional criticisms of binary distinctions and rather look at the details of how games are played. Finally, the paper proposes an alternative metaphor for game-playing, the puzzle piece.
Jesper Juul has convincingly argued that the conflict over the proper object of study has shifted from “rules or story” to “player or game.” But a key component of digital games is still missing from either of these oppositions: that of the computer itself. This paper offers a way of thinking about the phenomenology of the videogame from the perspective of the computer rather than the game or the player.
One of the informal properties often used to describe a new virtual world is its degree of openness. Yet what is an “open” virtual world? Does the phrase mean generally the same thing to different people? What distinguishes an open world from a less open world? Why does openness matter anyway? The answers to these questions cast light on an important, but shadowy, and uneasy, topic for virtual worlds: the relationship between those who construct the virtual, and those who use these constructions.
In this paper, we present a finite-state approach to constituency and therewith an analysis of coordination phenomena involving so-called non-constituents. We show that non-constituents can be seen as parts of fully-fledged constituents and therefore be coordinated in the same way. We have implemented an algorithm based on finite state automata that generates an LFG grammar assigning valid analyses to non-constituent coordination structures in the German language.
Generalized Two-Level Grammar (GTWOL) provides a new method for compilation of parallel replacement rules into transducers. The current paper identifies the role of generalized lenient composition (GLC) in this method. Thanks to the GLC operation, the compilation method becomes bipartite and easily extendible to capture various application modes. In the light of three notions of obligatoriness, a modification to the compilation method is proposed. We argue that the bipartite design makes implementation of parallel obligatoriness, directionality, length and rank based application modes extremely easy, which is the main result of the paper.
Morphological analyses based on word syntax approaches can encounter difficulties with long distance dependencies. The reason is that in some cases an affix has to have access to the inner structure of the form with which it combines. One solution is the percolation of features from ther inner morphemes to the outer morphemes with some process of feature unification. However, the obstacle of percolation constraints or stipulated features has lead some linguists to argue in favour of other frameworks such as, e.g., realizational morphology or parallel approaches like optimality theory. This paper proposes a linguistic analysis of two long distance dependencies in the morphology of Russian verbs, namely secondary imperfectivization and deverbal nominalization.We show how these processes can be reanalysed as local dependencies. Although finitestate frameworks are not bound by such linguistically motivated considerations, we present an implementation of our analysis as proposed in [1] that does not complicate the grammar or enlarge the network unproportionally.
The emergence of information extraction (IE) oriented pattern engines has been observed during the last decade. Most of them exploit heavily finite-state devices. This paper introduces ExPRESS – a new extraction pattern engine, whose rules are regular expressions over flat feature structures. The underlying pattern language is a blend of two previously introduced IE oriented pattern formalisms, namely, JAPE, used in the widely known GATE system, and the unificationbased XTDL formalism used in SProUT. A brief and technical overview of ExPRESS, its pattern language and the pool of its native linguistic components is given. Furthermore, the implementation of the grammar interpreter is addressed too.
In this work an extension of CSSR algorithm using Maximum Entropy Models is introduced. Preliminary experiments to perform Named Entity Recognition with this new system are presented.
In the last years, statistical machine translation has already demonstrated its usefulness within a wide variety of translation applications. In this line, phrase-based alignment models have become the reference to follow in order to build competitive systems. Finite state models are always an interesting framework because there are well-known efficient algorithms for their representation and manipulation. This document is a contribution to the evolution of finite state models towards a phrase-based approach. The inference of stochastic transducers that are based on bilingual phrases is carefully analysed from a finite state point of view. Indeed, the algorithmic phenomena that have to be taken into account in order to deal with such phrase-based finite state models when in decoding time are also in-depth detailed.
Temporal propositions are mapped to sets of strings that witness (in a precise sense) the propositions over discrete linear Kripke frames. The strings are collected into regular languages to ensure the decidability of entailments given by inclusions between languages. (Various notions of bounded entailment are shown to be expressible as language inclusions.) The languages unwind computations implicit in the logical (and temporal) connectives via a system of finite-state constraints adapted from finite-state morphology. Applications to Hybrid Logic and non-monotonic inertial reasoning are briefly considered.
This paper presents a system for the detection and correction of syntactic errors. It combines a robust morphosyntactic analyser and two groups of finite-state transducers specified using the Xerox Finite State Tool (xfst). One of the groups is used for the description of syntactic error patterns while the second one is used for the correction of the detected errors. The system has been tested on a corpus of real texts, containing both correct and incorrect sentences, with good results.
This paper describes the key aspects of the system SynCoP (Syntactic Constraint Parser) developed at the Berlin-Brandenburgische Akademie der Wissenschaften. The parser allows to combine syntactic tagging and chunking by means of constraint grammar using weighted finite state transducers (WFST). Chunks are interpreted as local dependency structures within syntactic tagging. The linguistic theories are formulated by criteria which are formalized by a semiring; these criteria allow structural preferences and gradual grammaticality. The parser is essentially a cascade of WFSTs. To find the most likely syntactic readings a best-path search is used.
We present an algorithm that computes a function that assigns consecutive integers to trees recognized by a deterministic, acyclic, finite-state, bottom-up tree automaton. Such function is called minimal perfect hashing. It can be used to identify trees recognized by the automaton. Its value may be seen as an index in some other data structures. We also present an algorithm for inverted hashing.
We introduce and discuss a number of issues that arise in the process of building a finite-state morphological analyzer for Urdu, in particular issues with potential ambiguity and non-concatenative morphology. Our approach allows for an underlyingly similar treatment of both Urdu and Hindi via a cascade of finite-state transducers that transliterates the very different scripts into a common ASCII transcription system. As this transliteration system is based on the XFST tools that the Urdu/Hindi common morphological analyzer is also implemented in, no compatibility problems arise.
Finite state methods for natural language processing often require the construction and the intersection of several automata. In this paper, we investigate the question of determining the best order in which these intersections should be performed. We take as an example lexical disambiguation in polarity grammars. We show that there is no efficient way to minimize the state complexity of these intersections.
Since Harris’ parser in the late 50s, multiword units have been progressively integrated in parsers. Nevertheless, in the most part, they are still restricted to compound words, that are more stable and less numerous. Actually, language is full of semi-fixed expressions that also form basic semantic units: semi-fixed adverbial expressions (e.g. time), collocations. Like compounds, the identification of these structures limits the combinatorial complexity induced by lexical ambiguity. In this paper, we detail an experiment that largely integrates these notions in a finite-state procedure of segmentation into super-chunks, preliminary to a parser.We show that the chunker, developped for French, reaches 92.9% precision and 98.7% recall. Moreover, multiword units realize 36.6% of the attachments within nominal and prepositional phrases.
This paper describes a two-level formalism where feature structures are used in contextual rules. Whereas usual two-level grammars describe rational sets over symbol pairs, this new formalism uses tree structured regular expressions. They allow an explicit and precise definition of the scope of feature structures. A given surface form may be described using several feature structures. Feature unification is expressed in contextual rules using variables, like in a unification grammar. Grammars are compiled in finite state multi-tape transducers.
This article describes a HMM-based word-alignment method that can selectively enforce a contiguity constraint. This method has a direct application in the extraction of a bilingual terminological lexicon from a parallel corpus, but can also be used as a preliminary step for the extraction of phrase pairs in a Phrase-Based Statistical Machine Translation system. Contiguous source words composing terms are aligned to contiguous target language words. The HMM is transformed into a Weighted Finite State Transducer (WFST) and contiguity constraints are enforced by specific multi-tape WFSTs. The proposed method is especially suited when basic linguistic resources (morphological analyzer, part-of-speech taggers and term extractors) are available for the source language only.
Nested complementation plays an important role in expressing counter- i.e. star-free and first-order definable languages and their hierarchies. In addition, methods that compile phonological rules into finite-state networks use double-nested complementation or “double negation”. This paper reviews how the double-nested complementation extends to a relatively new operation, generalized restriction (GR), coined by the author (Yli-Jyrä and Koskenniemi 2004). This operation encapsulates a double-nested complementation and elimination of a concatenation marker, diamond, whose finite occurrences align concatenations in the arguments of the operation. The paper demonstrates that the GR operation has an interesting potential in expressing regular languages, various kinds of grammars, bimorphisms and relations. This motivates a further study of optimized implementation of the operator.
Observational evidence exists that winds of massive stars are clumped. Many massive star systems are known as non-thermal particle production sites, as indicated by their synchrotron emission in the radio band. As a consequence they are also considered as candidate sites for non-thermal high-energy photon production up to gamma-ray energies. The present work considers the effects of wind clumpiness expected on the emitting relativistic particle spectrum in colliding wind systems, built up from the pool of thermal wind particles through diffusive particle acceleration, and taking into account inverse Compton and synchrotron losses. In comparison to a homogeneous wind, a clumpy wind causes flux variations of the emitting particle spectrum when the clump enters the wind collision region. It is found that the spectral features associated with this variability moves temporally from low to high energy bands with the time shift between any two spectral bands being dependent on clump size, filling factor, and the energy-dependence of particle energy gains and losses.
The most massive stars are those with the shortest but most active life. One group of massive stars, the Luminous Blue Variables (LBVs), of which only a few objects are known, are in particular of interest concerning the stability of stars. They have a high mass loss rate and are close to being instable. This is even more likely as rotation becomes an important factor in stellar evolution of these stars. Through massive stellar winds and sometimes giant eruptions, LBV nebulae are formed. Various aspects in the evolution in the LBV phase lead, beside the large scale morphological and kinematical differences, to a diversity of small structures like clumps, rims, and outflows in these nebulae.
Gamma-rays can be produced by the interaction of a relativistic jet and the matter of the stellar wind in the subclass of massive X-ray binaries known as “microquasars”. The relativistic jet is ejected from the surroundings of the compact object and interacts with cold protons from the stellar wind, producing pions that then quickly decay into gamma-rays. Since the resulting gamma-ray emissivity depends on the target density, the detection of rapid variability in microquasars with GLAST and the new generation of Cherenkov imaging arrays could be used to probe the clumped structure of the stellar wind. In particular, we show here that the relative fluctuation in gamma rays may scale with the square root of the ratio of porosity length to binary separation, $\sqrt{h/a}$, implying for example a ca. 10% variation in gamma ray emission for a quite moderate porosity, h/a ∼ 0.01.
The optical spectrum of Eta Carinae (η Car) is prominent in H I, He i and Fe ii wind lines, all of which vary both in absorption and emission with phase. The phase dependance is a consequence of the interaction between the two objects in the η Car binary (η Car A & B). The binary system is enshrouded by ejecta from previous mass ejection events and consequently, η Car B is not directly observable. We have traced the He i lines over η Car’s spectroscopic period, using HST/STIS data obtained with medium spectral, but high angular, resolving power, and created a radial velocity curve for the system. The He I lines are formed in the core of the system, and appear to be a composite of multiple features formed in spatially separated regions. The sources of their irregular line profiles are still not fully understood, but can be attributed to emission/absorption near the wind-wind interface and/or a direct consequence of the η Car A’s, massive, clumpy wind. This paper will discuss the spectral variability, the narrow emission structure of the He i lines and how clumpiness of the winds may impede the construction of the reliable radial velocity curve, necessary for characterizations of especially η Car B.
The spatially-resolved winds of the massive binary, Eta Carinae, extend an arcsecond on the sky, well beyond the 10 to 20 milliarcsecond binary orbital dimension. Stellar wind line profiles, observed at very different angular resolutions of VLTI/AMBER, HST/STIS and VLT/UVES, provide spatial information on the extended wind interaction structure as it changes with orbital phase. These same wind lines, observable in the starlight scattered off the foreground lobe of the dusty Homunculus, provide time-variant line profiles viewed from significantly different angles. Comparisons of direct and scattered wind profiles observed in the same epoch and at different orbital phases provide insight on the extended wind structure and promise the potential for three-dimensional imaging of the outer wind structures. Massive, long-lasting clumps, including the nebularWeigelt blobs, originated during the two historical ejection events. Wind interactions with these clumps are quite noticeable in spatially-resolved spectroscopy. As the 2009.0 minimum approaches, analysis of existing spectra and 3-D modeling are providing bases for key observations to gain further understanding of this complex massive binary.
The H.E.S.S. collaboration recently reported the discovery of VHE γ-ray emission coincident with the young stellar cluster Westerlund 2. This system is known to host a population of hot, massive stars, and, most particularly, the WR binary WR 20a. Particle acceleration to TeV energies in Westerlund 2 can be accomplished in several alternative scenarios, therefore we only discuss energetic constraints based on the total available kinetic energy in the system, the actual mass loss rates of respective cluster members, and implied gamma-ray production from processes such as inverse Compton scattering or neutral pion decay. From the inferred gammaray luminosity of the order of 1035erg/s, implications for the efficiency of converting available kinetic energy into non-thermal radiation associated with stellar winds in the Westerlund 2 cluster are discussed under consideration of either the presence or absence of wind clumping.
We model the line profile variability (lpv) in spectra of clumped stellar atmospheres using the Stochastic Clump Model (SCM) of the winds of early-type stars. In this model the formation of dense inhomogeneities (clumps) in the line driven winds is considered as being a stochastic process. It is supposed that the emission due to clumps mainly contributes to the intensities of emission lines in the stellar spectra. It is shown that in the framework of the SCM it is possible to reproduce both the mean line profiles and a common pattern of the lpv.
We study the time variability of emission lines in three WNE stars : WR 2 (WN2), WR 3 (WN3ha) and WR152 (WN3). While WR 2 shows no variability above the noise level, the other stars do show variation, which are like other WR stars in WR 152 but very fast in WR 3. From these motions, we deduce a value of β ∼1 for WR 3 that is like that seen in O stars and β ∼2–3 for WR 152, that is intermediate between other WR stars and WR 3.
Luminous Blue Variables show strong changes in their stellar wind on time scales of typically years to decades when they expand and contract radially at approximately constant luminosity. Micro-variability on shorter time scales and amplitudes can be observed superimposed to the larger scale radial changes. I will show long-term time series of high resolution spectra which we have collected in the past 20 years for many of the well known LBVs together with a few time series of weekly sampling (HR Car, R40, R71, R110, R127, S Dor) covering a time windows of up to a few months. Wind variability is seen on short and intermediate time scales with the line profiles changing from P Cygni to inverse P Cygni and double peeked profiles sometimes for the same star and spectral line. On longer time scales the ionisation levels for all chemical elements change drastically due to the strong change of the temperature on the stellar surface. While on the long term the characteristic radial changes may have impact on the over all mass loss rates, the variabilities and asymmetries on short and intermediate time scales may cause false estimates of the mass loss rates when confronting models with the observed line profiles
Hα observations of Rigel obtained on 184 nights during the past ten years with the 1-m telescope and ´echelle spectrograph of Ritter Observatory are surveyed. The line profiles were classified in terms of morphology. About 1/4 of them are of P Cygni type, about 15% inverse P Cygni, about 25% double-peaked, about 1/3 pure absorption, and a few are single emission lines. Transformation of the profile from one type to another typically takes a few days. Although the line stays in absorption for extended intervals, only one high-velocity absorption event of the intensity reported by Kaufer et al. (1996a) was observed, in late 2006. Late in this event, Hα absorption occurred farther to the red than the red wing of a plausible photospheric absorption component, an indication of infalling material. In general, as the absorption events come to an end, the emission typically returns with an inverse P Cygni profile. The Hα profile class shows no obvious correlation with the radial velocity of C II λ6578, a photospheric absorption line.
X-ray spectroscopy is a sensitive probe of stellar winds. X-rays originate from optically thin shock-heated plasma deep inside the wind and propagate outwards throughout absorbing cool material. Recent analyses of the line ratios from He-like ions in the X-ray spectra of O-stars highlighted problems with this general paradigm: the measured line ratios of highest ions are consistent with the location of the hottest X-ray emitting plasma very close to the base of the wind, perhaps indicating the presence of a corona, while measurements from lower ions conform with the wind-embedded shock model. Generally, to correctly model the emerging Xray spectra, a detailed knowledge of the cool wind opacities based on stellar atmosphere models is prerequisite. A nearly grey stellar wind opacity for the X-rays is deduced from the analyses of high-resolution X-ray spectra. This indicates that the stellar winds are strongly clumped. Furthermore, the nearly symmetric shape of X-ray emission line profiles can be explained if the wind clumps are radially compressed. In massive binaries the orbital variations of X-ray emission allow to probe the opacity of the stellar wind; results support the picture of strong wind clumping. In high-mass X-ray binaries, the stochastic X-ray variability and the extend of the stellar-wind part photoionized by X-rays provide further strong evidence that stellar winds consist of dense clumps.
By quantitatively fitting simple emission line profile models that include both atomic opacity and porosity to the Chandra X-ray spectrum of ζ Pup, we are able to explore the trade-offs between reduced mass-loss rates and wind porosity. We find that reducing the mass-loss rate of ζ Pup by roughly a factor of four, to 1.5 × 10−6 M⊙ yr−1, enables simple non-porous wind models to provide good fits to the data. If, on the other hand, we take the literature mass-loss rate of 6×10−6 M⊙ yr−1, then to produce X-ray line profiles that fit the data, extreme porosity lengths – of h∞ ≈ 3 R∗ – are required. Moreover, these porous models do not provide better fits to the data than the non-porous, low optical depth models. Additionally, such huge porosity lengths do not seem realistic in light of 2-D numerical simulations of the wind instability.
We review the effects of clumping on the profiles of resonance doublets. By allowing the ratio of the doublet oscillator strenghts to be a free parameter, we demonstrate that doublet profiles contain more information than is normally utilized. In clumped (or porous) winds, this ratio can lies between unity and the ratio of the f-values, and can change as a function of velocity and time, depending on the fraction of the stellar disk that is covered by material moving at a particular velocity at a given moment. Using these insights, we present the results of SEI modeling of a sample of B supergiants, ζ Pup and a time series for a star whose terminal velocity is low enough to make the components of its Si VIλλ1400 independent. These results are interpreted within the framewrok of the Oskinova et al. (2007) model, and demonstrate how the doublet profiles can be used to extract infromation about wind structure.
We present XMM-Newton Reflection Grating Spectrometer observations of pairs of X-ray emission line profiles from the O star ζ Pup that originate from the same He-like ion. The two profiles in each pair have different shapes and cannot both be consistently fit by models assuming the same wind parameters. We show that the differences in profile shape can be accounted for in a model including the effects of resonance scattering, which affects the resonance line in the pair but not the intercombination line. This implies that resonance scattering is also important in single resonance lines, where its effect is difficult to distinguish from a low effective continuum optical depth in the wind. Thus, resonance scattering may help reconcile X-ray line profile shapes with literature mass-loss rates.
We summarize Chandra observations of the emission line profiles from 17 OB stars. The lines tend to be broad and unshifted. The forbidden/intercombination line ratios arising from Helium-like ions provide radial distance information for the X-ray emission sources, while the H-like to He-like line ratios provide X-ray temperatures, and thus also source temperature versus radius distributions. OB stars usually show power law differential emission measure distributions versus temperature. In models of bow shocks, we find a power law differential emission measure, a wide range of ion stages, and the bow shock flow around the clumps provides transverse velocities comparable to HWHM values. We find that the bow shock results for the line profile properties, consistent with the observations of X-ray line emission for a broad range of OB star properties.
We present one-dimensional, time-dependent models of the clumps generated by the linedeshadowing instability. In order to follow the clumps out to distances of more than 1000 R∗, we use an efficient moving-box technique. We show that, within the approximations, the wind can remain clumped well into the formation region of the radio continuum.
INTEGRAL tripled the number of super-giant high-mass X-ray binaries (sgHMXB) known in the Galaxy by revealing absorbed and fast transient (SFXT) systems. Quantitative constraints on the wind clumping of massive stars can be obtained from the study of the hard X-ray variability of SFXT. A large fraction of the hard X-ray emission is emitted in the form of flares with a typical duration of 3 ksec, frequency of 7 days and luminosity of $10^{36}$ erg/s. Such flares are most probably emitted by the interaction of a compact object orbiting at $\sim10~R_*$ with wind clumps ($10^{22 ... 23}$ g) representing a large fraction of the stellar mass-loss rate. The density ratio between the clumps and the inter-clump medium is $10^{2 ... 4}$. The parameters of the clumps and of the inter-clump medium, derived from the SFXT flaring behavior, are in good agreement with macro-clumping scenario and line-driven instability simulations. SFXT are likely to have larger orbital radius than classical sgHMXB.
Magnetic fields influence the dynamics of hot-star winds and create large scale structure. Based on numerical magnetohydrodynamic (MHD) simulations, we model the wind of θ¹ Ori C, and then use the SEI method to compute synthetic line profiles for a range of viewing angles as function of rotational phase. The resulting dynamic spectrum for a moderately strong line shows a distinct modulation, but with a phase that seems at odds with available observations.
Discussion : X-rays
(2007)
Dynamical simulation of the “velocity-porosity” reduction in observed strength of stellar wind lines
(2007)
I use dynamical simulations of the line-driven instability to examine the potential role of the resulting flow structure in reducing the observed strength of wind absorption lines. Instead of the porosity length formalism used to model effects on continuum absorption, I suggest reductions in line strength can be better characterized in terms of a velocity clumping factor that is insensitive to spatial scales. Examples of dynamic spectra computed directly from instability simulations do exhibit a net reduction in absorption, but only at a modest 10-20% level that is well short of the ca. factor 10 required by recent analyses of PV lines.
The James Webb Space Telescope (JWST) is a large, infrared-optimized space telescope scheduled for launch in 2013. JWST will find the first stars and galaxies that formed in the early universe, connecting the Big Bang to our own Milky Way galaxy. JWST will peer through dusty clouds to see stars forming planetary systems, connecting the MilkyWay to our own Solar System. JWST’s instruments are designed to work primarily in the infrared range of 1 - 28 μm, with some capability in the visible range. JWST will have a large mirror, 6.5 m in diameter, and will be diffraction-limited at 2 μm (0.1 arcsec resolution). JWST will be placed in an L2 orbit about 1.5 million km from the Earth. The instruments will provide imaging, coronography, and multi-object and integral-field spectroscopy across the 1 - 28 μm wavelength range. The breakthrough capabilities of JWST will enable new studies of massive star winds from the Milky Way to the early universe.
Clumps in hot star winds can originate from shock compression due to the line driven instability. One-dimensional hydrodynamic simulations reveal a radial wind structure consisting of highly compressed shells separated by voids, and colliding with fast clouds. Two-dimensional simulations are still largely missing, despite first attempts. Clumpiness dramatically affects the radiative transfer and thus all wind diagnostics in the UV, optical, and in X-rays. The microturbulence approximation applied hitherto is currently superseded by a more sophisticated radiative transfer in stochastic media. Besides clumps, i.e. jumps in the density stratification, so-called kinks in the velocity law, i.e. jumps in dv/dr, play an eminent role in hot star winds. Kinks are a new type of radiative-acoustic shock, and propagate at super-Abbottic speed.
General Discussion
(2007)
We study the influence of clumping on the predicted wind structure of O-type stars. For this purpose we artificially include clumping into our stationary wind models. When the clumps are assumed to be optically thin, the radiative line force increases compared to corresponding unclumped models, with a similar effect on either the mass-loss rate or the terminal velocity (depending on the onset of clumping). Optically thick clumps, alternatively, might be able to decrease the radiative force.
We present the results of Monte Carlo mass-loss predictions for massive stars covering a wide range of stellar parameters. We critically test our predictions against a range of observed massloss rates – in light of the recent discussions on wind clumping. We also present a model to compute the clumping-induced polarimetric variability of hot stars and we compare this with observations of Luminous Blue Variables, for which polarimetric variability is larger than for O and Wolf-Rayet stars. Luminous Blue Variables comprise an ideal testbed for studies of wind clumping and wind geometry, as well as for wind strength calculations, and we propose they may be direct supernova progenitors.
Many hot stars exhibit stochastic polarimetric variability, thought to arise from clumping low in the wind. Here we investigate the wind properties required to reproduce this variability using analytic models, with particular emphasis on Luminous Blue Variables. We find that the winds must be highly structured, consisting of a large number of optically-thin clumps; while we find that the overall level of polarization should scale with mass-loss rate – consistent with observations of LBVs. The models also predict variability on very short timescales, which is supported by the results of a recent polarimetric monitoring campaign.
We investigate the effect of wind clumping on the dynamics of Wolf-Rayet winds, by means of the Potsdam Wolf-Rayet (PoWR) hydrodynamic atmosphere models. In the limit of microclumping the radiative acceleration is generally enhanced. We examine the reasons for this effect and show that the resulting wind structure depends critically on the assumed radial dependence of the clumping factor D(r). The observed terminal wind velocities for WR stars imply that D(r) increases to very large values in the outer part of the wind, in agreement with the assumption of detached expanding shells.
Overwhelming observational and theoretical evidence suggests that the winds of massive stars are highly clumped. We briefly discuss the influence of clumping on model diagnostics and the difficulties of allowing for the influence of clumping on model spectra. Because of its simplicity, and because of computational ease, most spectroscopic analyses incorporate clumping using the volume filling factor. The biases introduced by this approach are uncertain. To investigate alternative clumping models, and to help determine the validity of parameters derived using the volume filling factor method, we discuss results derived using an alternative model in which we assume that the wind is composed of optically thick shells.
We report FUSE observations in 2005–2006 of three O-type, double-lined spectroscopic binaries in the Magellanic Clouds. The systems have very short periods (1.4–2.25 d), represent rare, young evolutionary stages of massive stars and binaries, and provide a unique glimpse at some of the most massive systems that form in dense clusters of massive stars. Improved orbit parameters, including revised masses, for LH54-425 are derived from new ctio spectroscopy. The systems are: LH54-425 in the LMC (O3V + O5V, P=2.25d, 62+37M⊙), J053441-693139 in the LMC (O2-3If+O6V, P=1.4 d, 41+27M⊙), and Hodge 53-47 in the SMC (O6V + O4-5IIIf, P=2.2 d, 24+14M⊙, where the O4 star appears to be less massive than the O6 star). Their short periods indicates that wind interaction and mass transfer are likely important factors in their evolution. The spectra provide quantitative and systematic studies of phase-dependent stellar wind properties, wind collision effects in O+O binaries at lower metallicities, improved radial velocity curves, and FUV spectro-photometric changes as a function of orbital phase.
We present preliminary results of a tailored atmosphere analysis of six Galactic WC stars using UV, optical, and mid-infrared Spitzer IRS data. With these data, we are able to sample regions from 10 to 10³ stellar radii, thus to determine wind clumping in different parts of the wind. Ultimately, derived wind parameters will be used to accuratelymeasure neon abundances, and to so test predicted nuclear-reaction rates.
Mass accretion onto compact objects through accretion disks is a common phenomenon in the universe. It is seen in all energy domains from active galactic nuclei through cataclysmic variables (CVs) to young stellar objects. Because CVs are fairly easy to observe, they provide an ideal opportunity to study accretion disks in great detail and thus help us to understand accretion also in other energy ranges. Mass accretion in these objects is often accompanied by mass outflow from the disks. This accretion disk wind, at least in CVs, is thought to be radiatively driven, similar to O star winds. WOMPAT, a 3-D Monte Carlo radiative transfer code for accretion disk winds of CVs is presented.
We apply the 3-dimensional radiative transport codeWind3D to 3D hydrodynamic models of Corotating Interaction Regions to fit the detailed variability of Discrete Absorption Components observed in Si iv UV resonance lines of HD 64760 (B0.5 Ib). We discuss important effects of the hydrodynamic input parameters on these large-scale equatorial wind structures that determine the detailed morphology of the DACs computed with 3D transfer. The best fit model reveals that the CIR in HD 64760 is produced by a source at the base of the wind that lags behind the stellar surface rotation. The non-corotating coherent wind structure is an extended density wave produced by a local increase of only 0.6% in the smooth symmetric wind mass-loss rate.
Modeling expanding atmospheres is a difficult task because of the extreme non-LTE situation, the need to account for complex model atoms, especially for the iron-group elements with their millions of lines, and because of the supersonic expansion. Adequate codes have been developed e.g. by Hillier (CMFGEN), the Munich group (Puls, Pauldrach), and in Potsdam (PoWR code, Hamann et al.). While early work was based on the assumption of a smooth and homogeneous spherical stellar wind, the need to account for clumping became obvious about ten years ago. A relatively simple first-order clumping correction was readily implemented into the model codes. However, its simplifying assumptions are severe. Most importantly, the clumps are taken to be optically thin at all frequencies (”microclumping”). We discuss the consequences of this approximation and describe an approach to account for optically thick clumps (“macroclumping”). First results demonstrate that macroclumping can generally reduce the strength of spectral features, depending on their optical thickness. The recently reported discrepancy between the Hα diagnostic and the Pv resonance lines in O star spectra can be resolved without decreasing the mass-loss rates, when macroclumping is taken into account.
Clumping in Galactic WN stars : a comparison of mass loss rates from UV/optical & radio diagnostics
(2007)
The mass loss rates and other parameters for a large sample of Galactic WN stars have been revised by Hamann et al. (2006), using the most up-to date Potsdam Wolf-Rayet (PoWR) model atmospheres. For a sub-sample of these stars exist measurements of their radio free-free emission. After harmonizing the adopted distance and terminal wind velocities, we compare the mass loss rates obtained from the two diagnostics. The differences are discussed as a possible consequence of different clumping contrast in the line-forming and radio-emitting regions.
Recent studies of massive O-type stars present clear evidences of inhomogeneous and clumped winds. O-type (H-rich) central stars of planetary nebulae (CSPNs) are in some ways the low mass–low luminosity analogous of those massive stars. In this contribution, we present preliminary results of our on-going multi-wavelength (FUV, UV and optical) study of the winds of Galactic CSPNs. Particular emphasis will be given to the clumping factors derived by means of optical lines (Hα and Heii 4686) and “classic” FUV (and UV) lines.
We exploit time-series $FUSE$ spectroscopy to {\it uniquely} probe spatial structure and clumping in the fast wind of the central star of the H-rich planetary nebula NGC~6543 (HD~164963). Episodic and recurrent optical depth enhancements are discovered in the P{\sc v} absorption troughs, with some evidence for a $\sim$ 0.17-day modulation time-scale. The characteristics of these features are essentially identical to the discrete absorption components' (DACs) commonly seen in the UV lines of massive OB stars, suggesting the temporal structures seen in NGC~6543 likely have a physical origin that is similar to that operating in massive, luminous stars. The mechanism for forming coherent perturbations in the outflows is therefore apparently operating equally in the radiation-pressure-driven winds of widely differing momenta ($\mdot$$v_\infty$$R_\star^{0.5}$) and flow times, as represented by OB stars and CSPN.
This paper outlines a newly-developed method to include the effects of time variability in the radiative transfer code CMFGEN. It is shown that the flow timescale is often large compared to the variability timescale of LBVs. Thus, time-dependent effects significantly change the velocity law and density structure of the wind, affecting the derivation of the mass-loss rate, volume filling factor, wind terminal velocity, and luminosity. The results of this work are directly applicable to all active LBVs in the Galaxy and in the LMC, such as AG Car, HR Car, S Dor and R 127, and could result in a revision of stellar and wind parameters. The massloss rate evolution of AG Car during the last 20 years is presented, highlighting the need for time-dependent models to correctly interpret the evolution of LBVs.
We discuss the results of time-resolved spectroscopy of three presumably single Population I Wolf-Rayet stars in the Small Magellanic Cloud, where the ambient metallicity is $\sim 1/5 Z_\odot$. We were able to detect and follow numerous small-scale wind-embedded inhomogeneities in all observed stars. The general properties of the moving features, such as their velocity dispersions, emissivities and average accelerations, closely match the corresponding characteristics of small-scale inhomogeneities in the winds of Galactic Wolf-Rayet stars.
The influence of the wind to the total continuum of OB supergiants is discussed. For wind velocity distributions with β > 1.0, the wind can have strong influence to the total continuum emission, even at optical wavelengths. Comparing the continuum emission of clumped and unclumped winds, especially for stars with high β values, delivers flux differences of up to 30% with maximum in the near-IR. Continuum observations at these wavelengths are therefore an ideal tool to discriminate between clumped and unclumped winds of OB supergiants.
Massive stars usually form groups such as OB associations. Their fast stellar winds sweep up collectively the surrounding insterstellar medium (ISM) to generate superbubbles. Observations suggest that superbubble evolution on the surrounding ISM can be very irregular. Numerical simulations considering these conditions could help to understand the evolution of these superbubbles and to clarify the dynamics of these objects as well as the difference between observed X-ray luminosities and the predicted ones by the standard model (Weaver et al. 1977).
We present the latest results on the observational dependence of the mass-loss rate in stellar winds of O and early-B stars on the metal content of their atmospheres, and compare these with predictions. Absolute empirical rates for the mass loss of stars brighter than 10$^{5.2} L_{\odot}$, based on H$\alpha$ and ultraviolet (UV) wind lines, are found to be about a factor of two higher than predictions. If this difference is attributed to inhomogeneities in the wind this would imply that luminous O and early-B stars have clumping factors in their H$\alpha$ and UV line forming regime of about a factor of 3--5. The investigated stars cover a metallicity range $Z$ from 0.2 to 1 $Z_{\odot}$. We find a hint towards smaller clumping factors for lower $Z$. The derived clumping factors, however, presuppose that clumping does not impact the predictions of the mass-loss rate. We discuss this assumption and explain how we intend to investigate its validity in more detail.