Refine
Year of publication
Document Type
- Conference Proceeding (584) (remove)
Language
- English (493)
- German (78)
- Multiple languages (10)
- French (2)
- Russian (1)
Keywords
- social media (5)
- Archiv (4)
- Information Structure (4)
- Nachlass (4)
- COVID-19 (3)
- Cloud Computing (3)
- E-Learning (3)
- E-Mail Tracking (3)
- ERP (3)
- MOOC (3)
Institute
- Extern (139)
- Fachgruppe Betriebswirtschaftslehre (78)
- Institut für Biochemie und Biologie (58)
- Department Sport- und Gesundheitswissenschaften (43)
- Institut für Ernährungswissenschaft (38)
- Department Psychologie (30)
- Institut für Künste und Medien (23)
- Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung (23)
- Institut für Informatik und Computational Science (20)
- Institut für Slavistik (17)
- Institut für Chemie (15)
- Institut für Physik und Astronomie (15)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (14)
- Institut für Geowissenschaften (13)
- Wirtschaftswissenschaften (10)
- Institut für Mathematik (9)
- Bürgerliches Recht (7)
- Historisches Institut (7)
- Sonderforschungsbereich 632 - Informationsstruktur (7)
- Wirtschafts- und Sozialwissenschaftliche Fakultät (7)
- Institut für Germanistik (6)
- MenschenRechtsZentrum (6)
- Strukturbereich Kognitionswissenschaften (6)
- Department Linguistik (5)
- Institut für Umweltwissenschaften und Geographie (5)
- Hasso-Plattner-Institut für Digital Engineering GmbH (4)
- Fachgruppe Politik- & Verwaltungswissenschaft (3)
- Fachgruppe Soziologie (3)
- Institut für Jüdische Studien und Religionswissenschaft (3)
- Institut für Romanistik (3)
- Öffentliches Recht (3)
- Fakultät für Gesundheitswissenschaften (2)
- Institut für Philosophie (2)
- Kommunalwissenschaftliches Institut (2)
- Digital Engineering Fakultät (1)
- Hochschulambulanz (1)
- Juristische Fakultät (1)
- Lehreinheit für Wirtschafts-Arbeit-Technik (1)
- Philosophische Fakultät (1)
- Sozialwissenschaften (1)
- Theodor-Fontane-Archiv (1)
- Weitere Einrichtungen (1)
- Zentrum für Lehrerbildung und Bildungsforschung (ZeLB) (1)
Many hot stars exhibit stochastic polarimetric variability, thought to arise from clumping low in the wind. Here we investigate the wind properties required to reproduce this variability using analytic models, with particular emphasis on Luminous Blue Variables. We find that the winds must be highly structured, consisting of a large number of optically-thin clumps; while we find that the overall level of polarization should scale with mass-loss rate – consistent with observations of LBVs. The models also predict variability on very short timescales, which is supported by the results of a recent polarimetric monitoring campaign.
We investigate the effect of wind clumping on the dynamics of Wolf-Rayet winds, by means of the Potsdam Wolf-Rayet (PoWR) hydrodynamic atmosphere models. In the limit of microclumping the radiative acceleration is generally enhanced. We examine the reasons for this effect and show that the resulting wind structure depends critically on the assumed radial dependence of the clumping factor D(r). The observed terminal wind velocities for WR stars imply that D(r) increases to very large values in the outer part of the wind, in agreement with the assumption of detached expanding shells.
Overwhelming observational and theoretical evidence suggests that the winds of massive stars are highly clumped. We briefly discuss the influence of clumping on model diagnostics and the difficulties of allowing for the influence of clumping on model spectra. Because of its simplicity, and because of computational ease, most spectroscopic analyses incorporate clumping using the volume filling factor. The biases introduced by this approach are uncertain. To investigate alternative clumping models, and to help determine the validity of parameters derived using the volume filling factor method, we discuss results derived using an alternative model in which we assume that the wind is composed of optically thick shells.
We report FUSE observations in 2005–2006 of three O-type, double-lined spectroscopic binaries in the Magellanic Clouds. The systems have very short periods (1.4–2.25 d), represent rare, young evolutionary stages of massive stars and binaries, and provide a unique glimpse at some of the most massive systems that form in dense clusters of massive stars. Improved orbit parameters, including revised masses, for LH54-425 are derived from new ctio spectroscopy. The systems are: LH54-425 in the LMC (O3V + O5V, P=2.25d, 62+37M⊙), J053441-693139 in the LMC (O2-3If+O6V, P=1.4 d, 41+27M⊙), and Hodge 53-47 in the SMC (O6V + O4-5IIIf, P=2.2 d, 24+14M⊙, where the O4 star appears to be less massive than the O6 star). Their short periods indicates that wind interaction and mass transfer are likely important factors in their evolution. The spectra provide quantitative and systematic studies of phase-dependent stellar wind properties, wind collision effects in O+O binaries at lower metallicities, improved radial velocity curves, and FUV spectro-photometric changes as a function of orbital phase.
We present preliminary results of a tailored atmosphere analysis of six Galactic WC stars using UV, optical, and mid-infrared Spitzer IRS data. With these data, we are able to sample regions from 10 to 10³ stellar radii, thus to determine wind clumping in different parts of the wind. Ultimately, derived wind parameters will be used to accuratelymeasure neon abundances, and to so test predicted nuclear-reaction rates.
Mass accretion onto compact objects through accretion disks is a common phenomenon in the universe. It is seen in all energy domains from active galactic nuclei through cataclysmic variables (CVs) to young stellar objects. Because CVs are fairly easy to observe, they provide an ideal opportunity to study accretion disks in great detail and thus help us to understand accretion also in other energy ranges. Mass accretion in these objects is often accompanied by mass outflow from the disks. This accretion disk wind, at least in CVs, is thought to be radiatively driven, similar to O star winds. WOMPAT, a 3-D Monte Carlo radiative transfer code for accretion disk winds of CVs is presented.
We apply the 3-dimensional radiative transport codeWind3D to 3D hydrodynamic models of Corotating Interaction Regions to fit the detailed variability of Discrete Absorption Components observed in Si iv UV resonance lines of HD 64760 (B0.5 Ib). We discuss important effects of the hydrodynamic input parameters on these large-scale equatorial wind structures that determine the detailed morphology of the DACs computed with 3D transfer. The best fit model reveals that the CIR in HD 64760 is produced by a source at the base of the wind that lags behind the stellar surface rotation. The non-corotating coherent wind structure is an extended density wave produced by a local increase of only 0.6% in the smooth symmetric wind mass-loss rate.
Modeling expanding atmospheres is a difficult task because of the extreme non-LTE situation, the need to account for complex model atoms, especially for the iron-group elements with their millions of lines, and because of the supersonic expansion. Adequate codes have been developed e.g. by Hillier (CMFGEN), the Munich group (Puls, Pauldrach), and in Potsdam (PoWR code, Hamann et al.). While early work was based on the assumption of a smooth and homogeneous spherical stellar wind, the need to account for clumping became obvious about ten years ago. A relatively simple first-order clumping correction was readily implemented into the model codes. However, its simplifying assumptions are severe. Most importantly, the clumps are taken to be optically thin at all frequencies (”microclumping”). We discuss the consequences of this approximation and describe an approach to account for optically thick clumps (“macroclumping”). First results demonstrate that macroclumping can generally reduce the strength of spectral features, depending on their optical thickness. The recently reported discrepancy between the Hα diagnostic and the Pv resonance lines in O star spectra can be resolved without decreasing the mass-loss rates, when macroclumping is taken into account.
Clumping in Galactic WN stars : a comparison of mass loss rates from UV/optical & radio diagnostics
(2007)
The mass loss rates and other parameters for a large sample of Galactic WN stars have been revised by Hamann et al. (2006), using the most up-to date Potsdam Wolf-Rayet (PoWR) model atmospheres. For a sub-sample of these stars exist measurements of their radio free-free emission. After harmonizing the adopted distance and terminal wind velocities, we compare the mass loss rates obtained from the two diagnostics. The differences are discussed as a possible consequence of different clumping contrast in the line-forming and radio-emitting regions.
Recent studies of massive O-type stars present clear evidences of inhomogeneous and clumped winds. O-type (H-rich) central stars of planetary nebulae (CSPNs) are in some ways the low mass–low luminosity analogous of those massive stars. In this contribution, we present preliminary results of our on-going multi-wavelength (FUV, UV and optical) study of the winds of Galactic CSPNs. Particular emphasis will be given to the clumping factors derived by means of optical lines (Hα and Heii 4686) and “classic” FUV (and UV) lines.
We exploit time-series $FUSE$ spectroscopy to {\it uniquely} probe spatial structure and clumping in the fast wind of the central star of the H-rich planetary nebula NGC~6543 (HD~164963). Episodic and recurrent optical depth enhancements are discovered in the P{\sc v} absorption troughs, with some evidence for a $\sim$ 0.17-day modulation time-scale. The characteristics of these features are essentially identical to the discrete absorption components' (DACs) commonly seen in the UV lines of massive OB stars, suggesting the temporal structures seen in NGC~6543 likely have a physical origin that is similar to that operating in massive, luminous stars. The mechanism for forming coherent perturbations in the outflows is therefore apparently operating equally in the radiation-pressure-driven winds of widely differing momenta ($\mdot$$v_\infty$$R_\star^{0.5}$) and flow times, as represented by OB stars and CSPN.
This paper outlines a newly-developed method to include the effects of time variability in the radiative transfer code CMFGEN. It is shown that the flow timescale is often large compared to the variability timescale of LBVs. Thus, time-dependent effects significantly change the velocity law and density structure of the wind, affecting the derivation of the mass-loss rate, volume filling factor, wind terminal velocity, and luminosity. The results of this work are directly applicable to all active LBVs in the Galaxy and in the LMC, such as AG Car, HR Car, S Dor and R 127, and could result in a revision of stellar and wind parameters. The massloss rate evolution of AG Car during the last 20 years is presented, highlighting the need for time-dependent models to correctly interpret the evolution of LBVs.
We discuss the results of time-resolved spectroscopy of three presumably single Population I Wolf-Rayet stars in the Small Magellanic Cloud, where the ambient metallicity is $\sim 1/5 Z_\odot$. We were able to detect and follow numerous small-scale wind-embedded inhomogeneities in all observed stars. The general properties of the moving features, such as their velocity dispersions, emissivities and average accelerations, closely match the corresponding characteristics of small-scale inhomogeneities in the winds of Galactic Wolf-Rayet stars.
The influence of the wind to the total continuum of OB supergiants is discussed. For wind velocity distributions with β > 1.0, the wind can have strong influence to the total continuum emission, even at optical wavelengths. Comparing the continuum emission of clumped and unclumped winds, especially for stars with high β values, delivers flux differences of up to 30% with maximum in the near-IR. Continuum observations at these wavelengths are therefore an ideal tool to discriminate between clumped and unclumped winds of OB supergiants.
Massive stars usually form groups such as OB associations. Their fast stellar winds sweep up collectively the surrounding insterstellar medium (ISM) to generate superbubbles. Observations suggest that superbubble evolution on the surrounding ISM can be very irregular. Numerical simulations considering these conditions could help to understand the evolution of these superbubbles and to clarify the dynamics of these objects as well as the difference between observed X-ray luminosities and the predicted ones by the standard model (Weaver et al. 1977).
We present the latest results on the observational dependence of the mass-loss rate in stellar winds of O and early-B stars on the metal content of their atmospheres, and compare these with predictions. Absolute empirical rates for the mass loss of stars brighter than 10$^{5.2} L_{\odot}$, based on H$\alpha$ and ultraviolet (UV) wind lines, are found to be about a factor of two higher than predictions. If this difference is attributed to inhomogeneities in the wind this would imply that luminous O and early-B stars have clumping factors in their H$\alpha$ and UV line forming regime of about a factor of 3--5. The investigated stars cover a metallicity range $Z$ from 0.2 to 1 $Z_{\odot}$. We find a hint towards smaller clumping factors for lower $Z$. The derived clumping factors, however, presuppose that clumping does not impact the predictions of the mass-loss rate. We discuss this assumption and explain how we intend to investigate its validity in more detail.
While there is strong evidence for clumping in the winds of massive hot stars, very little is known about clumping in the winds from Central Stars. We have checked [WC]-type CSPN winds for clumping by inspecting the electron-scattering line wings. At least for three stars we found indications for wind inhomogeneities.
We report on new mass-loss rate estimates for O stars in six massive binaries using the amplitude of orbital-phase dependent, linear-polarimetric variability caused by electron scattering off free electrons in the winds. Our estimated mass-loss rates for luminous O stars are independent of clumping. They suggest similar clumping corrections as for WR stars and do not support the recently proposed reduction in mass-loss rates of O stars by one or two orders of magnitude.
Clumping in O-star winds
(2007)
We have analyzed the spectra of seven Galactic O4 supergiants, with the NLTE wind code CMFGEN. For all stars, we have found that clumped wind models match well lines from different species spanning a wavelength range from FUV to optical, and remain consistent with Hα data. We have achieved an excellent match of the P V λλ1118, 1128 resonance doublet and N IV λ1718, as well as He II λ4686 suggesting that our physical description of clumping is adequate. We find very small volume filling factors and that clumping starts deep in the wind, near the sonic point. The most crucial consequence of our analysis is that the mass loss rates of O stars need to be revised downward significantly, by a factor of 3 and more compared to those obtained from smooth-wind models.
I discuss observational evidence – independent of the direct spectral diagnostics of stellar winds themselves – suggesting that mass-loss rates for O stars need to be revised downward by roughly a factor of three or more, in line with recent observed mass-loss rates for clumped winds. These independent constraints include the large observed mass-loss rates in LBV eruptions, the large masses of evolved massive stars like LBVs and WNH stars, WR stars in lower metallicity environments, observed rotation rates of massive stars at different metallicity, supernovae that seem to defy expectations of high mass-loss rates in stellar evolution, and other clues. I pay particular attention to the role of feedback that would result from higher mass-loss rates, driving the star to the Eddington limit too soon, and therefore making higher rates appear highly implausible. Some of these arguments by themselves may have more than one interpretation, but together they paint a consistent picture that steady line-driven winds of O-type stars have lower mass-loss rates and are significantly clumped.
The P v λλ1118, 1128 resonance doublet is an extraordinarily useful diagnostic of O-star winds, because it bypasses the traditional problems associated with determining mass-loss rates from UV resonance lines. We discuss critically the assumptions and uncertainties involved with using P v to diagnose mass-loss rates, and conclude that the large discrepancies between massloss rates determined from P v and the rates determined from “density squared” emission processes pose a significant challenge to the “standard model” of hot-star winds. The disparate measurements can be reconciled if the winds of O-type stars are strongly clumped on small spatial scales, which in turn implies that mass-loss rates based on Hα or radio emission are too large by up to an order of magnitude.
In the old days (pre ∼1990) hot stellar winds were assumed to be smooth, which made life fairly easy and bothered no one. Then after suspicious behaviour had been revealed, e.g. stochastic temporal variability in broadband polarimetry of single hot stars, it took the emerging CCD technology developed in the preceding decades (∼1970-80’s) to reveal that these winds were far from smooth. It was mainly high-S/N, time-dependent spectroscopy of strong optical recombination emission lines in WR, and also a few OB and other stars with strong hot winds, that indicated all hot stellar winds likely to be pervaded by thousands of multiscale (compressible supersonic turbulent?) structures, whose driver is probably some kind of radiative instability. Quantitative estimates of clumping-independent mass-loss rates came from various fronts, mainly dependent directly on density (e.g. electron-scattering wings of emission lines, UV spectroscopy of weak resonance lines, and binary-star properties including orbital-period changes, electron-scattering, and X-ray fluxes from colliding winds) rather than the more common, easier-to-obtain but clumping-dependent density-squared diagnostics (e.g. free-free emission in the IR/radio and recombination lines, of which the favourite has always been Hα). Many big questions still remain, such as: What do the clumps really look like? Do clumping properties change as one recedes from the mother star? Is clumping universal? Does the relative clumping correction depend on $\dot{M}$ itself?
Mass loss is a very important aspect of the life of massive stars. After briefly reviewing its importance, we discuss the impact of the recently proposed downward revision of mass loss rates due to clumping (difficulty to form Wolf-Rayet stars and production of critically rotating stars). Although a small reduction might be allowed, large reduction factors around ten are disfavoured. We then discuss the possibility of significant mass loss at very low metallicity due to stars reaching break-up velocities and especially due to the metal enrichment of the surface of the star via rotational and convective mixing. This significant mass loss may help the first very massive stars avoid the fate of pair-creation supernova, the chemical signature of which is not observed in extremely metal poor stars. The chemical composition of the very low metallicity winds is very similar to that of the most metal poor star known to date, HE1327-2326 and offer an interesting explanation for the origin of the metals in this star. We also discuss the importance of mass loss in the context of long and soft gamma-ray bursts and pair-creation supernovae. Finally, we would like to stress that mass loss in cooler parts of the HR-diagram (luminous blue variable and yellow and red supergiant stages) are much more uncertain than in the hot part. More work needs to be done in these areas to better constrain the evolution of the most massive stars.
Stellar winds play an important role for the evolution of massive stars and their cosmic environment. Multiple lines of evidence, coming from spectroscopy, polarimetry, variability, stellar ejecta, and hydrodynamic modeling, suggest that stellar winds are non-stationary and inhomogeneous. This is referred to as 'wind clumping'. The urgent need to understand this phenomenon is boosted by its far-reaching implications. Most importantly, all techniques to derive empirical mass-loss rates are more or less corrupted by wind clumping. Consequently, mass-loss rates are extremely uncertain. Within their range of uncertainty, completely different scenarios for the evolution of massive stars are obtained. Settling these questions for Galactic OB, LBV and Wolf-Rayet stars is prerequisite to understanding stellar clusters and galaxies, or predicting the properties of first-generation stars. In order to develop a consistent picture and understanding of clumped stellar winds, an international workshop on 'Clumping in Hot Star Winds' was held in Potsdam, Germany, from 18. - 22. June 2007. About 60 participants, comprising almost all leading experts in the field, gathered for one week of extensive exchange and discussion. The Scientific Organizing Committee (SOC) included John Brown (Glasgow), Joseph Cassinelli (Madison), Paul Crowther (Sheffield), Alex Fullerton (Baltimore), Wolf-Rainer Hamann (Potsdam, chair), Anthony Moffat (Montreal), Stan Owocki (Newark), and Joachim Puls (Munich). These proceedings contain the invited and contributed talks presented at the workshop, and document the extensive discussions.
During the last few years there was a tremendous growth of scientific activities in the fields related to both Physics and Control theory: nonlinear dynamics, micro- and nanotechnologies, self-organization and complexity, etc. New horizons were opened and new exciting applications emerged. Experts with different backgrounds starting to work together need more opportunities for information exchange to improve mutual understanding and cooperation. The Conference "Physics and Control 2007" is the third international conference focusing on the borderland between Physics and Control with emphasis on both theory and applications. With its 2007 address at Potsdam, Germany, the conference is located for the first time outside of Russia. The major goal of the Conference is to bring together researchers from different scientific communities and to gain some general and unified perspectives in the studies of controlled systems in physics, engineering, chemistry, biology and other natural sciences. We hope that the Conference helps experts in control theory to get acquainted with new interesting problems, and helps experts in physics and related fields to know more about ideas and tools from the modern control theory.
brandial06 was the tenth in a series of workshops that aims to bring together researchers working on the semantics and pragmatics of dialogues in fields such as artificial intelligence, formal semantics and pragmatics, computational linguistics, philosophy, and psychology. This volume collects all presented papers and posters and gives abstracts of the invited talks.
On the basis of the Dynamic Syntax framework, this paper argues that the production pressures in dialogue determining alignment effects and given versus new informational effects also drive the shift from case-rich free word order systems without clitic pronouns into systems with clitic pronouns with rigid relative ordering. The paper introduces assumptions of Dynamic Syntax, in particular the building up of interpretation through structural underspecification and update, sketches the attendant account of production with close coordination of parsing and production strategies, and shows how what was at the Latin stage a purely pragmatic, production-driven decision about linear ordering becomes encoded in the clitics in theMedieval Spanish system which then through successive steps of routinization yield the modern systems with immediately pre-verbal fixed clitic templates.
We analyze anaphoric phenomena in the context of building an input understanding component for a conversational system for tutoring mathematics. In this paper, we report the results of data analysis of two sets of corpora of dialogs on mathematical theorem proving. We exemplify anaphoric phenomena, identify factors relevant to anaphora resolution in our domain and extensions to the input interpretation component to support it.
Received views of utterance context in pragmatic theory characterize the occurrent subjective states of interlocutors using notions like common knowledge or mutual belief. We argue that these views are not compatible with the uncertainty and robustness of context-dependence in humanhuman dialogue. We present an alternative characterization of utterance context as objective and normative. This view reconciles the need for uncertainty with received intuitions about coordination and meaning in context, and can directly inform computational approaches to dialogue.
Goal-oriented dialog as a collaborative subordinated activity involving collective acceptance
(2006)
Modeling dialog as a collaborative activity consists notably in specifying the contain of the Conversational Common Ground and the kind of social mental state involved. In previous work (Saget, 2006), we claim that Collective Acceptance is the proper social attitude for modeling Conversational Common Ground in the particular case of goal-oriented dialog. We provide a formalization of Collective Acceptance, besides elements in order to integrate this attitude in a rational model of dialog are provided; and finally, a model of referential acts as being part of a collaborative activity is provided. The particular case of reference has been chosen in order to exemplify our claims.
A key problem for models of dialogue is to explain the mechanisms involved in generating and responding to clarification requests. We report a 'Maze task' experiment that investigates the effect of 'spoof' clarification requests on the development of semantic co-ordination. The results provide evidence of both local and global semantic co-ordination phenomena that are not captured by existing dialogue co-ordination models.
How does a shared lexicon arise in population of agents with differing lexicons, and how can this shared lexicon be maintained over multiple generations? In order to get some insight into these questions we present an ALife model in which the lexicon dynamics of populations that possess and lack metacommunicative interaction (MCI) capabilities are compared. We ran a series of experiments on multi-generational populations whose initial state involved agents possessing distinct lexicons. These experiments reveal some clear differences in the lexicon dynamics of populations that acquire words solely by introspection contrasted with populations that learn using MCI or using a mixed strategy of introspection and MCI. The lexicon diverges at a faster rate for an introspective population, eventually collapsing to one single form which is associated with all meanings. This contrasts sharply with MCI capable populations in which a lexicon is maintained, where every meaning is associated with a unique word. We also investigated the effect of increasing the meaning space and showed that it speeds up the lexicon divergence for all populations irrespective of their acquisition method.
Demonstratives, in particular gestures that "only" accompany speech, are not a big issue in current theories of grammar. If we deal with gestures, fixing their function is one big problem, the other one is how to integrate the representations originating from different channels and, ultimately, how to determine their composite meanings. The growing interest in multi-modal settings, computer simulations, human-machine interfaces and VRapplications increases the need for theories ofmultimodal structures and events. In our workshopcontribution we focus on the integration of multimodal contents and investigate different approaches dealing with this problem such as Johnston et al. (1997) and Johnston (1998), Johnston and Bangalore (2000), Chierchia (1995), Asher (2005), and Rieser (2005).
A new method is used in an eye-tracking pilot experiment which shows that it is possible to detect differences in common ground associated with the use of minimally different types of indefinite anaphora. Following Richardson and Dale (2005), cross recurrence quantification analysis (CRQA) was used to show that the tandem eye movements of two Swedish-speaking interlocutors are slightly more coupled when they are using fully anaphoric indefinite expressions than when they are using less anaphoric indefinites. This shows the potential of CRQA to detect even subtle processing differences in ongoing discourse.
Verbal or visual? : How information is distributed across speech and gesture in spatial dialog
(2006)
In spatial dialog like in direction giving humans make frequent use of speechaccompanying gestures. Some gestures convey largely the same information as speech while others complement speech. This paper reports a study on how speakers distribute meaning across speech and gesture, and depending on what factors. Utterance meaning and the wider dialog context were tested by statistically analyzing a corpus of direction-giving dialogs. Problems of speech production (as indicated by discourse markers and disfluencies), the communicative goals, and the information status were found to be influential, while feedback signals by the addressee do not have any influence.
We describe an experiment to gather original data on geometrical aspects of pointing. In particular, we are focusing upon the concept of the pointing cone, a geometrical model of a pointing’s extension. In our setting we employed methodological and technical procedures of a new type to integrate data from annotations as well as from tracker recordings. We combined exact information on position and orientation with rater’s classifications. Our first results seem to challenge classical linguistic and philosophical theories of demonstration in that they advise to separate pointings from reference.
Classical SDRT (Asher and Lascarides, 2003) discussed essential features of dialogue like adjacency pairs or corrections and up-dating. Recent work in SDRT (Asher, 2002, 2005) aims at the description of natural dialogue. We use this work to model situated communication, i.e. dialogue, in which sub-sentential utterances and gestures (pointing and grasping) are used as conventional modes of communication. We show that in addition to cognitive modelling in SDRT, capturing mental states and speech-act related goals, special postulates are needed to extract meaning out of contexts. Gestural meaning anchors Discourse Referents in contextually given domains. Both sorts of meaning are fused with the meaning of fragments to get at fully developed dialogue moves. This task accomplished, the standard SDRT machinery, tagged SDRSs, rhetorical relations, the up-date mechanism, and the Maximize Discourse Coherence constraint generate coherent structures. In sum, meanings from different verbal and non-verbal sources are assembled using extended SDRT to form coherent wholes.
We present a formal analysis of iconic coverbal gesture. Our model describes the incomplete meaning of gesture that’s derivable from its form, and the pragmatic reasoning that yields a more specific interpretation. Our formalism builds on established models of discourse interpretation to capture key insights from the descriptive literature on gesture: synchronous speech and gesture express a single thought, but while the form of iconic gesture is an important clue to its interpretation, the content of gesture can be resolved only by linking it to its context.
In two experiments, many annotators marked antecedents for discourse deixis as unconstrained regions of text. The experiments show that annotators do converge on the identity of these text regions, though much of what they do can be captured by a simple model. Demonstrative pronouns are more likely than definite descriptions to be marked with discourse antecedents. We suggest that our methodology is suitable for the systematic study of discourse deixis.
We present a new analysis of illocutionary forces in dialogue. We analyze them as complex conversational moves involving two dimensions: what Speaker commits herself to and what she calls on Addressee to perform. We start from the analysis of speech acts such as confirmation requests or whimperatives, and extend the analysis to seemingly simple speech acts, such as statements and queries. Then, we show how to integrate our proposal in the framework of the Grammar for Conversation (Ginzburg, to app.), which is adequate for modelling agents' information states and how they get updated.
Claiming that cross-speaker "but" can signal correction in dialogue, we start by describing the types of corrections "but" can communicate by focusing on the Speech Act (SA) communicated in the previous turn and address the ways in which "but" can correct what is communicated. We address whether "but" corrects the proposition, the direct SA or the discourse relation communicated in the previous turn. We will also briefly address other relations signalled by cross-turn "but". After presenting a typology of the situations "but" can correct, we will address how these corrections can be modelled in the Information State model of dialogue, motivating this work by showing how it can be used to potentially avoid misunderstandings. We wrap up by showing how the model presented here updates beliefs in the Information State representation of the dialogue and can be used to facilitate response deliberation.
An account is presented of the focus properties, common ground effect and dialogue behaviour of the accented German discourse marker "doch" and the accented sentence negation "nicht". It is argued that "doch" and "nicht" evoke as a focus alternative the logical complement of the proposition expressed by the sentence in which they occur, and that an analysis in terms of contrastive focus accounts for their effect on the common ground and their function in dialogue.
We present an analysis of student language input in a corpus of tutoring dialogue in the domain of symbolic differentiation. Our focus on procedural tutoring makes the dialogue comparable to collaborative problem-solving (CPS). Existing CPS models describe the process of negotiating plans and goals, which also fits procedural tutoring. However, we provide a classification of student utterances and corpus annotation which shows that approximately 28% of non-trivial student language in this corpus is not accounted for by existing models, and addresses other functions, such as evaluating past actions or correcting mistakes. Our analysis can be used as a foundation for improving models of tutoring dialogue.
We present an extension to a comprehensive context model that has been successfully employed in a number of practical conversational dialogue systems. The model supports the task of multimodal fusion as well as that of reference resolution in a uniform manner. Our extension consists of integrating implicitly mentioned concepts into the context model and we show how they serve as candidates for reference resolution.
In semi-arid savannas, unsustainable land use can lead to degradation of entire landscapes, e.g. in the form of shrub encroachment. This leads to habitat loss and is assumed to reduce species diversity. In BIOTA phase 1, we investigated the effects of land use on population dynamics on farm scale. In phase 2 we scale up to consider the whole regional landscape consisting of a diverse mosaic of farms with different historic and present land use intensities. This mosaic creates a heterogeneous, dynamic pattern of structural diversity at a large spatial scale. Understanding how the region-wide dynamic land use pattern affects the abundance of animal and plant species requires the integration of processes on large as well as on small spatial scales. In our multidisciplinary approach, we integrate information from remote sensing, genetic and ecological field studies as well as small scale process models in a dynamic region-wide simulation tool. <hr> Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006.
Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Decisions for the conservation of biodiversity and sustainable management of natural resources are typically related to large scales, i.e. the landscape level. However, understanding and predicting the effects of land use and climate change on scales relevant for decision-making requires to include both, large scale vegetation dynamics and small scale processes, such as soil-plant interactions. Integrating the results of multiple BIOTA subprojects enabled us to include necessary data of soil science, botany, socio-economics and remote sensing into a high resolution, process-based and spatially-explicit model. Using an example from a sustainably-used research farm and a communally used and degraded farming area in semiarid southern Namibia we show the power of simulation models as a tool to integrate processes across disciplines and scales.
Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Mit der politischen Wende in den Staaten des ehemaligen Ostblockes wurde für viele militärisch genutzte Flächen ein tiefgreifender Nutzungswandel eingeleitet. Truppenübungsplätze als stark gestörte Bestandteile unserer Kulturlandschaft weisen auf großen Flächen naturschutzfachlich wertvolle Habitatmosaike mit speziellen Lebensgemeinschaften auf. Der Nutzungswandel ist mit einer Veränderung der Vegetationsstrukturen (Sukzession) und weiteren landschaftsökologischen Prozessen verbunden. Der ehemalige Truppenübungsplatz Döberitz im Norden der Landeshauptstadt Potsdam kann auf eine lange militärische Nutzungsgeschichte verweisen (erste Manöver des Soldatenkönigs im Jahr 1713). Nach 1992 wurden das NSG Döberitzer Heide (3.415 ha) und das NSG Ferbitzer Bruch (1.155 ha) ausgewiesen. Als Schutzgebiete nach der Vogelschutzrichtlinie sind sie Bestandteile des kohärenten Schutzgebietssystems Natura 2000 der europäischen Gemeinschaft. Trotz des Schutzstatus und der militärischen Altlasten unterliegt das Gebiet als größte zusammenhängende Naturfläche im engeren Verflechtungsraum des Landes Brandenburg einem hohen Nutzungsdruck. <hr> Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
In ihrem Bemühen, landwirtschaftliche Flächen standortgerecht zu bewirtschaften, sammelt eine zunehmende Anzahl landwirtschaftlicher Betriebe Informationen über die räumlich-zeitliche Verteilung von Boden- und Pflanzenmerkmalen auf ihren Schlägen. Diese Informationen dienen unmittelbar (Echtzeitansatz) oder mittelbar (Kartenansatz) zur Dosierung von Dünge- und Pflanzenschutzmitteln (Präzise Landbewirtschaftung). Zur Datensammlung werden vorrangig fahrzeuggestützte Sensoren und VIS- und NIR-Luftbilder, aufgenommen aus Sportflugzeugen, verwendet. Erste Betriebe erwerben von Dienstleistungsunternehmen aufbereitete Satelliten-Fernerkundungsdaten. Die landwirtschaftliche und agrartechnische Forschung ist bestrebt, die grundlegenden Muster (z.B. des Ertragspotentials) zu erkennen und damit den Aufwand der Betriebe für eine regelmäßige Informationserfassung gering zu halten. <hr> Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Fluvial systems are one of the major features shaping a landscape. They adjust to the prevailing tectonic and climatic setting and therefore are very sensitive markers of changes in these systems. If their response to tectonic and climatic forcing is quantified and if the climatic signal is excluded, it is possible to derive a local deformation history. Here, we investigate fluvial terraces and erosional surfaces in the southern Chilean forearc to assess a long-term geomorphic and hence tectonic evolution. Remote sensing and field studies of the Nahuelbuta Range show that the long-term deformation of the Chilean forearc is manifested by breaks in topography, sequences of differentially uplifted marine, alluvial and strath terraces as well as tectonically modified river courses and drainage basins. We used SRTM-90-data as basic elevation information for extracting and delineating drainage networks. We calculated hypsometric curves as an indicator for basin uplift, stream-length gradient indices to identify stream segments with anomalous slopes, and longitudinal river profiles as well as DS-plots to identify knickpoints and other anomalies. In addition, we investigated topography with elevation-slope graphs, profiles, and DEMs to reveal erosional surfaces. During the first field trip we already measured palaeoflow directions, performed pebble counting and sampled the fluvial terraces in order to apply cosmogenic nuclide dating (<sup>10Be, <sup>26Al) as well as provenance analyses. Our preliminary analysis of the Coastal Cordillera indicates a clear segmentation between the northern and southern parts of the Nahuelbuta Range. The Lanalhue Fault, a NW-SE striking fault zone oblique to the plate boundary, defines the segment boundary. Furthermore, we find a complex drainage re-organisation including a drainage reversal and wind gap on the divide between the Tirúa and Pellahuén basins east of the town Tirúa. The coastal basins lost most of their Andean sediment supply areas that existed in Tertiary and in part during early Pleistocene time. Between the Bío-Bío and Imperial rivers no Andean river is recently capable to traverse the Coastal Cordillera, suggesting ongoing Quaternary uplift of the entire range. From the spatial distribution of geomorphic surfaces in this region two uplift signals may be derived: (1) a long-term differential uplift process, active since the Miocene and possibly caused by underplating of subducted trench sediments, (2) a younger, local uplift affecting only the northern part of the Nahuelbuta Range that may be caused by the interaction of the forearc with the subduction of the Mocha Fracture Zone at the latitude of the Arauco peninsula. Our approach thus provides results in our attempt to decipher the characteristics of forearc development of active convergent margins using long-term geomorphic indicators. Furthermore, it is expected that our ongoing assessment will constrain repeatedly active zones of deformation. <hr> Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
The rigorous development, application and validation of distributed hydrological models obligates to evaluate data in a spatially distributed way. In particular, spatial model predictions such as the distribution of soil moisture, runoff generating areas or nutrient-contributing areas or erosion rates, are to be assessed against spatially distributed observations. Also model inputs, such as the distribution of modelling units derived by GIS and remote sensing analyses, should be evaluated against groundbased observations of landscape characteristics. So far, however, quantitative methods of spatial field comparison have rarely been used in hydrology. In this paper, we present algorithms that allow to compare observed and simulated spatial hydrological data. The methods can be applied for binary and categorical data on regular grids. They comprise cell-by-cell algorithms, cell-neighbourhood approaches that account for fuzziness of location, and multi-scale algorithms that evaluate the similarity of spatial fields with changing resolution. All methods provide a quantitative measure of the similarity of two maps. The comparison methods are applied in two mountainous catchments in southern Germany (Brugga, 40 km<sup>2) and Austria (Löhnersbach, 16 km<sup>2). As an example of binary hydrological data, the distribution of saturated areas is analyzed in both catchments. For categorical data, vegetation zones that are associated with different runoff generation mechanisms are analyzed in the Löhnersbach. Mapped spatial patterns are compared to simulated patterns from terrain index calculations and from satellite image analysis. It is discussed how particular features of visual similarity between the spatial fields are captured by the quantitative measures, leading to recommendations on suitable algorithms in the context of evaluating distributed hydrological models.
Abbildende Spektrometrie
(2006)
Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Integration of digital elevation models and satellite images to investigate geological processes.
(2006)
In order to better understand the geological boundary conditions for ongoing or past surface processes geologists face two important questions: 1) How can we gain additional knowledge about geological processes by analyzing digital elevation models (DEM) and satellite images and 2) Do these efforts present a viable approach for more efficient research. Here, we will present case studies at a variety of scales and levels of resolution to illustrate how we can substantially complement and enhance classical geological approaches with remote sensing techniques. Commonly, satellite and DEM based studies are being used in a first step of assessing areas of geologic interest. While in the past the analysis of satellite imagery (e.g. Landsat TM) and aerial photographs was carried out to characterize the regional geologic characteristics, particularly structure and lithology, geologists have increasingly ventured into a process-oriented approach. This entails assessing structures and geomorphic features with a concept that includes active tectonics or tectonic activity on time scales relevant to humans. In addition, these efforts involve analyzing and quantifying the processes acting at the surface by integrating different remote sensing and topographic data (e.g. SRTM-DEM, SSM/I, GPS, Landsat 7 ETM, Aster, Ikonos…). A combined structural and geomorphic study in the hyperarid Atacama desert demonstrates the use of satellite and digital elevation data for assessing geological structures formed by long-term (millions of years) feedback mechanisms between erosion and crustal bending (Zeilinger et al., 2005). The medium-term change of landscapes during hundred thousands to millions years in a more humid setting is shown in an example from southern Chile. Based on an analysis of rivers/watersheds combined with landscapes parameterization by using digital elevation models, the geomorphic evolution and change in drainage pattern in the coastal Cordillera can be quantified and put into the context of seismotectonic segmentation of a tectonically active region. This has far-reaching implications for earthquake rupture scenarios and hazard mitigation (K. Rehak, see poster on IMAF Workshop). Two examples illustrate short-term processes on decadal, centennial and millennial time scales: One study uses orogen scale precipitation gradients derived from remotely sensed passive microwave data (Bookhagen et al., 2005a). They demonstrate how debris flows were triggered as a response of slopes to abnormally strong rainfall in the interior parts of the Himalaya during intensified monsoons. The area of the orogen that receives high amounts of precipitation during intensified monsoons also constitutes numerous landslide deposits of up to 1km<sup>3 volume that were generated during intensified monsoon phase at about 27 and 9 ka (Bookhagen et al., 2005b). Another project in the Swiss Alps compared sets of aerial photographs recorded in different years. By calculating high resolution surfaces the mass transport in a landslide could be reconstructed (M. Schwab, Universität Bern). All these examples, although representing only a short and limited selection of projects using remote sense data in geology, have as a common approach the goal to quantify geological processes. With increasing data resolution and new sensors future projects will even enable us to recognize more patterns and / or structures indicative of geological processes in tectonically active areas. This is crucial for the analysis of natural hazards like earthquakes, tsunamis and landslides, as well as those hazards that are related to climatic variability. The integration of remotely sensed data at different spatial and temporal scales with field observations becomes increasingly important. Many of presently highly populated places and increasingly utilized regions are subject to significant environmental pressure and often constitute areas of concentrated economic value. Combined remote sensing and ground-truthing in these regions is particularly important as geologic, seismicity and hydrologic data may be limited here due to the recency of infrastructural development. Monitoring ongoing processes and evaluating the remotely sensed data in terms of recurrence of events will greatly enhance our ability to assess and mitigate natural hazards. <hr> Dokument 1: Foliensatz | Dokument 2: Abstract <hr> Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Die Präsentation gibt zuerst einen Überblick über mögliche Parameter für die Wasserhaushaltsmodellierung, die aus Fernerkundungs(FE)-daten generell abgeleitet werden können. Bei der Beschreibung der Ableitungsverfahren dieser Parameter aus (FE)-Daten wird auf die Landnutzung, Vegetationsindices und die reale Evapotranspiration (ETr) fokussiert. Die Verfahren zur Bestimmung der ETr aus optischen FE-Daten lassen grob wie folgt gliedern : • Direkte Ableitung der Evapotranspiration aus radiometrisch bestimmten Oberflächen-temperaturen • Ableitung von Modellinputdaten wie z.B. Globalstrahlung, Albedo, Blattflächeniondex LAI und NDVI aus FE-Daten zur Anwendung von SoilVegetation-AtmosphereTransfer- und Energiebilanzmodellen wie z.B. SEBAL (Bastiaansen et al . 1998) • Kombinierte Anwendung verschiedenster Sensoren wie SAR-ERS1, LANDSAT-TM, NOAA-AHVRR mit SVAT-Modellen und hydrologischen Einzugsgebietsmodellen Die Validierung dieser Methoden wurde in verschiedenen Messkampagnen wie z.B. Lo-trex10E-HIBE, FIFE oder HAPEX-Sahel durchgeführt. Dabei wurde die aus dem entspre-chenden Sensor abgeleitete ETr mit gemessenen ETr-Raten von Ankerstationen innerhalb eines definierten Gebietes verglichen. Diese Ankerstationen leiteten die ETr aus Profil-, Ed-dy-Flux-, oder Szintillometermessungen ab. Durchgängige längere Zeitreihen der ETr sind nur mit FE-Daten mit hoher Wiederholungsrate wie z.B. NOAA-AVHRR, MODIS hoher Zeitauflösung möglich Mit Landsat-TM z.B. ergeben sich dagegen nur „Snap Shots“ der ETr von einzelnen Tagen. Daher wurden oftmals Multisensorverfahren d.h. Kombination von z.B. Landsat-TM mit NOAA-AVHRR eingesetzt oder die FE-Daten nur für die Erhebung zeitin-varianter Eingangsdaten (z.B. Landnutzung) und zur raumbezogenen Validierung der ETr-Berechnungen von hydrologischen Modellen verwendet. Im zweiten Teil des Vortrags wird ein Anwendungsbeispiel für den Versuch einer räumliche Validierung eines Wasserhaus-haltsmodells über NDVI-ETr-Datenprodukte aus Landsat-TM5-Daten für das Stobbergebiet. Ein weiteres Anwendungsbeispiel für die Einbindung von Landnutzungsdatenprodukten aus Landsat-TM5-Daten in die Wasserhaushaltmodellierung für das Ucker-Einzugsgebiet schliesst den Vortrag ab. <hr> Dokument 1: Foliensatz | Dokument 2: Abstract <hr> Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Problemstellung • Geoökologische Prozessforschung versucht für große Landschaftsausschnitte, die in der Natur ablaufenden und vom Menschen beeinflussten Prozesse mit Hilfe von Modellen nachzuvollziehen • exakte Erfassung der Ausstattung des Untersuchungsraumes ist wesentliche Voraussetzung für eine wirklichkeitsnahe Abbildung • Modelle derzeit weder in der Lage, alle ablaufenden Prozesse in die Betrachtung einzubeziehen, noch präzise Eingangsdaten bei der Beschreibung des Ausgangszustandes zu verarbeiten • häufig liegen Modelleingangsdaten nicht in der notwendigen Präzision vor • In Modellen wird Ausstattung eines Untersuchungsgebietes über den Boden, den Grundwassereinfluss und die Flächennutzung beschrieben • Flächennutzung besitzt weitgehend statische Elemente (Nutzungstypen Wald, Gewässer, Siedlung) und hochdynamische Elemente (jährlicher Wechsel der Fruchtart auf jedem Acker) • Bedarf nach detaillierter (lage- und zeitkonkreter) Eingabe der Verteilung der Ackerfrüchte im Modellzeitraum, da Landwirtschaft als eine der bedeutenden Quellen für diffusen Nährstoffeintrag ins Ökosystem angesehen wird Stand der Forschung • bei Erfassung von Kulturen der Landwirtschaft aus Fernerkundungsdaten hat sich multitemporale Klassifizierung als sinnvoll erwiesen, weil sich anhand einer Einzelaufnahme die verschiedenen Kulturen nicht sicher trennen lassen • Klassifizierung erfolgt mit überwachten Methoden unter Verwendung von Trainingsflächen im Datensatz, von denen die dort angebaute Frucht bekannt ist • in die Klassifizierung werden zusätzliche Informationen einbezogen (Fuzzy), die Auskunft über die Wahrscheinlichkeit des Auftretens der Frucht geben (Anbaueignung in Abhängigkeit von Hangneigung, Niederschlag, Höhenlage, Boden) Die Ergebnisse dieser Klassifikationen sind meist nicht auf andere Landschaftsausschnitte und Anbaujahre übertragbar, weil die Ausprägung der Spektralsignatur einer Kultur durch veränderte Boden- und Witterungsbedingungen variiert. Lösungsansatz • auf Basis von Satellitendaten und Anbauinformationen aus 15 aufeinander folgenden Jahren (35 Aufnahmetermine) sollten von Witterung und Boden unabhängige Jahreskurven der spektralen Charakteristik wichtiger Ackerkulturen gewonnen werden, die den Wachstumsverlauf der Pflanzen beschreiben • diese Kurven sollen anstelle von Trainingsgebieten zur multitemporalen Klassifizierung von Daten eines Anbaujahres herangezogen werden Schlussfolgerungen und Ausblick • Prinzipiell erscheint Vorgehen erfolgreich, jedoch in Abhängigkeit von der Brauchbarkeit der herangezogenen Szenen schwankt Güte des Ergebnisses noch • Verfahren stellt wesentlichen Fortschritt zu bisherigem Vorgehen auf Trainingsflächenbasis dar • ist zumindest im Untersuchungsgebiet immer wieder ohne weitere Kenntnis von Anbauinformationen anwendbar, lediglich exakte phänologische Datierung der dann verwendeten Aufnahmen erforderlich • für andere Gebiete (Variation in Niederschlag und Boden) ist Anpassung der phänologischen Datierung der Kurven erforderlich (Form ist weiter verwendbar) • optimale Bildkombination zur Trennung aller Kulturen ist: Anfang/Mitte April – Mitte Mai – Anfang Juli – Mitte August – Mitte September • Kombination sollte bei verbesserter Verfügbarkeit von Daten beschaffbar sein • problematisch scheinen Trockensituationen im Mai und Juni zu sein, so dass zu schnell reifende Wintergetreide nicht richtig erkannt werden, Bedarf Bodeninformationen einzubeziehen • Trennung von Hackfrüchten weiterhin problematisch (wie schon in bisherigen Verfahren), führt zu übermäßigen Anteilen im Ergebnis, in Abhängigkeit vom Anbauanteil besser vernachlässigen • Einbeziehung von Fuzzyinformationen erscheint sinnvoll • Zusammenhang von Bodengüte und Frucht (Anbaueignung eines Bodens für eine Frucht) • Wasserverfügbarkeit am Standort (in Abhängigkeit von Speichervermögen des Bodens, Grundwasseranschluss und Niederschlag) • Summe der Niederschläge bis zum Aufnahmezeitpunkt (Trockenheitsindex) <hr> Dokument 1: Foliensatz | Dokument 2: Abstract <hr> Interdisziplinäres Zentrum für Musterdynamik und Angewandte Fernerkundung Workshop vom 9. - 10. Februar 2006
Stand des IMAF zu Beginn des Jahres 2006
Zum 1. April 2005 wurde per Beschluss des Rektorats der Universität Potsdam das Interdisziplinäre Zentrum für Musterdynamik und Angewandte Fernerkundung (IMAF) an der Universität Potsdam eingerichtet. Diesem Beschluss gingen knapp zwei Jahre konzeptionelle, organisatorische und administrative Vorarbeiten voraus. Inzwischen ist das IMAF also offiziell gegründet, der Vorstand wurde „bestellt“ (Prof. M. Mutti. Prof. E. Zehe, Prof. A. Bronstert), der Geschäftsführer bzw. wissenschaftliche Koordinator Dr. M. Kühling arbeitet in dieser Funktion seit Sommer 2005 und seit kurzem ist auch die 1. Version der Homepage des IMAF (http://www.uni-potsdam.de/imaf/) frei geschaltet. Auch die Infrastruktur des IMAF ist in der Entstehungsphase: Büroräume sind versprochen (wenn auch noch nicht bezugsfertig) im Haus 13 auf dem Campus Golm der Universität Potsdam und der 1. erfolgreiche Drittmittelantrag erbrachte 8 leistungsfähige Tischrechner und einen Server für das IMAF aus EU-Mitteln. Wichtiger als die administrativen und organisatorischen Arbeiten sind aber die inhaltlichen Forstschritte. Hier ist die große Resonanz, die die Gründung des IMAF sowohl innerhalb als auch außerhalb der Universität gefunden hat, besonders erfreulich. Über 30 Angehörige des Zentrums sind inzwischen zu verzeichnen und es gibt bereits eine Reihe von wissenschaftlichen Projektinitiativen und Ideen für dieses Zentrum. Neben den wissenschaftlichen Arbeiten am IMAF ist ein zweites Hauptziel für dieses Zentrum die Entwicklung und der Ausbau eines strukturierten Ausbildungsangebotes für Musterdynamik und angewandte Fernerkundung. Dies sollen gleichermaßen Masterstudenten als auch Doktoranden der Universität Potsdam und der mit ihr assoziierten außeruniversitären Institute nutzen. Zudem werden Kurse und Weiterbildungsveranstaltungen mit nationalen und internationalen Experten angestrebt. Neben diesen positiven Entwicklungen gibt es auch (noch ??) über einige Mängel zu berichten:
Das Sekretariat ist nach wie vor unbesetzt, die Finanzausstattung des Zentrums ist völlig ungenügend und die im Konzept für das Zentrum beantragte Wissenschaftlerstelle für Softwareanwendung ist nicht in Sicht. Für einen Erfolg des Zentrums ist es unbedingt notwendig, dass sich diese Situation deutlich verbessert!!
Forschungsschwerpunkte des IMAF
Räumliche Muster und deren Struktur in der Umwelt
Räumliche Muster sind in vielen naturwissenschaftlichen Disziplinen (Hydrologie, Ökologie, Geologie, Biologie, Chemie, Physik) von zentraler Bedeutung. Z.B. bestimmen die räumlichen (und zeitlichen) Muster von Bodeneigenschaften und Vegetation in ihrem Zusammenspiel mit den Mustern von Niederschlag und Strahlungsinput maßgeblich den Wasser- und Stoffhaushalt auf unterschiedlichsten Skalen und führen über Rückkopplung wiederum zu Veränderungen in Klima, Vegetation und Ökosystemen. Vom kleinräumigen Transport von Schadstoffen und von der Hochwasserentstehung bis zur Frage nach den regionalen und globalen Veränderungen von Klima, Vegetation und Landnutzung seien hier nur einige Problemkreise genannt, in denen Muster und Musterdynamik eine zentrale Stellung einnehmen. Darüber hinaus liefert die Betrachtung der zeitlichen Veränderung von räumlichen Mustern, in Ergänzung zur klassischen Erfassung dynamischer Prozesse in Form von Messungen lokaler zeitlicher Änderungen, eine völlig neue Perspektive auf Dynamik und eröffnet damit völlig neue wissenschaftliche Möglichkeiten. Aktuelle und sehr drängende Fragen innerhalb dieses Forschungsschwerpunktes sind unter anderem:
• Analyse der generelle Raumstruktur von Geodaten (Variabilität, Struktur, Konnektivität);
• Thematische Verbindungen verschiedener Datenebenen und Möglichkeiten für deren Assimilation;
• Möglichkeiten und Grenzen des Skalenübergangs zwischen verschiedenen räumlichen Auflösungen und Informationsquellen;
• Ableitung der zeitlichen Dynamik bzw. Entwicklung von großen flächenhaften Datenfeldern.
Angewandte Fernerkundung
Wie keine andere Technik bietet die Fernerkundung in jeglicher Form (unter anderem Satelliten, flugzeuggetragene Sensoren, Wetterradar und auch geophysikalische Methoden) umfangreiche Möglichkeiten, räumliche Muster und deren zeitliche Veränderungen zu erfassen. Allen Methoden der Fernerkundung gemein ist, dass sie nur indirekte Ergebnisse liefern. Das heißt, es besteht nur ein mittelbarer Zusammenhang zwischen dem beobachteten Signal, meist der Reflektivität oder Emissivität elektromagnetischer Strahlung in verschiedenen Spektralbereichen (optisch oder Radar), und der eigentlich interessierenden Größe, wie dem Feuchtezustand der Vegetation, der Bodenfeuchte oder Bodenrauhigkeit, der Niederschlagsintensität, dem Zustand der Schneedecke oder der Ausdehnung eines Oberflächenfilms auf Gewässern. Ein Satellitenbild enthält beispielsweise immer die spektrale Signatur des räumlichen Musters mehrerer der oben genannten Einflussgrößen, was die Extraktion oder Diskriminierung der eigentlich interessierenden Größe erschwert. Dieser „vermischte“ Charakter der Fernerkundungsdaten bietet aber auch immense Chancen. So lassen sich durch geeignete Interpretationsverfahren aus jedem mit hohem finanziellem und technischem Aufwand erstellten Satellitenbild zahlreiche und im Detail völlig unterschiedliche Fragestellungen bearbeiten. Die Extraktion der gewünschten Information aus dem Fernerkundungssignal führt mathematisch gesehen meist auf die Lösung so genannter inverser, schlecht gestellter Probleme. Somit beinhaltet die interdisziplinäre Nutzung von Fernerkundung auch ein hohes methodisches Synergiepotential. Durch die heutigen technischen Möglichkeiten zur Archivierung auch sehr umfangreicher raumbezogener Informationen ist die Bearbeitung zu jedem beliebigen Zeitpunkt nach der Aufnahme möglich – zum Beispiel bis entsprechend lange Zeitreihen und/oder geeignete Interpretationsverfahren zur Verfügung stehen. Tatsächlich dürfte der weitaus größte Teil der raumbezogenen Informationen, die in den bisher erhobenen Fernerkundungsdaten stecken, nur in Ansätzen ausgewertet sein. Einer bereits sehr hoch entwickelten technischen Dimension der Fernerkundung steht ein gewisses Defizit im Umfang ihrer Anwendung in den verschiedenen naturwissenschaftlichen Disziplinen gegenüber. Aktuelle und sehr drängende Fragen innerhalb dieses Forschungsschwerpunktes sind unter anderem:
• Nutzung der räumlichen und inhaltlichen Breite von Fernerkundungsinformationen;
• Verbindung mit automatisierten, u.a. geophysikalischen Methoden des „ground-truthings“;
• Identifizierung der Grenzen bzgl. Repräsentanz der Daten (spektral, raum-zeitliche Auflösung);
• Verbindung unterschiedlicher Methoden der Fernerkundung und der Geophysik.
Dieser Beitrag illustriert die o.g. Fragestellungen anhand einiger Darstellungen aus verschiedenen wissenschaftlichen Disziplinen und erläutert 2 Beispiele zu beabsichtigten Forschungsprojekten:
• Erfassung und Bedeutung von Boden-Oberflächeneigenschaften auf die Abflussbildung von Landschaften;
• Phänomene des Stofftransportes in homogenen vs. heterogenen Böden.
Die Banken- und Börsensysteme Europas befinden sich in einem grundlegenden Wandel. Globalisierungsprozesse der Finanzmärkte, die Harmonisierung der aufsichtsrechtlichen Rahmenbedingungen sowie neue Informationstechnologien stellen hohe Anforderungen. Die Integration Russlands in die Weltwirtschaft erfordert eine weitere Liberalisierung, Fortentwicklung und Anpassung der Banken- und Börsenstrukturen an internationale und europäische Standards, um die nationale und internationale Kapitalallokation effizienter zu gestalten. Finanz-, Banken- und Börsenexperten aus deutschen und russischen Universitäten lehren und forschen seit Jahren zu diesen Fragestellungen. Der Potsdamer Workshop stellt die Ergebnisse verschiedener Lehrstühle aus Potsdam, St. Petersburg und Moskau zur Diskussion. Mit folgenden Beiträgen: Detlev HUMMEL: Integration und Perspektiven des Europäischen Kapitalmarktes, Hans-Georg PETERSEN: Tax Competition, Tax Havens and Capital income Taxation, Yakov M. MIRKIN: Importance of the European Capital Market for Russia, Christoph LATTEMANN: Elektronische Systeme für den Wertpapierhandel in Europa und Entwicklungsmöglichkeiten für Russland – eine explorative Studie, Hans-Joachim MASSENBERG: Lage und Perspektiven des deutschen Bankensektors, Ralf WIEGERT: Russia’s Banking System, the Central Bank and the Exchange Rate Regime, Galina BELOGLAZOWA: Entwicklungsstrategie des russischen Bankensystems und die Rolle der Sberbank, Anatole KOSMATCHEV: Wettbewerb im Bankensystem Russlands, Tatjana NIKITINA: Auswirkungen von Basel II auf das russische Bankensystem, Julia PLAKITKINA: Auslandsbanken in Russland, Natalia DOUMNAYA: Russian-German Economic Relations under Globalization, Tatjana NIKOLAJEWA: Investitionen, Banken und Wirtschaftswachstum in Russland, Alexej LINKOV: Die aktuellen Probleme der Außenverschuldung Russlands
Nachlässe sind persönliches Eigentum und unterliegen deshalb keiner Abgabepflicht. Der Wunsch des Nachlassers bezüglich der weiteren Aufbewahrung seines schriftlichen Erbes ist demzufolge primär gegenüber allen unseren Wünschen. Wir können nicht fordern, sondern nur bitten, uns durch eigene Leistungen anbieten und die zukünftigen Nachlassenden oder deren Erben überzeugen. Die nicht vorhandene institutionelle Zuständigkeit für die Übernahme von Nachlässen erzeugt die Reibungspunkte zwischen den Einrichtungen, die sich um den Erwerb von Nachlässen bemühen: Archive – Bibliotheken – Museen - Sammlungen. Die Wünsche zum Erwerb des Nachlasses einer bestimmten Person – egal ob Wissenschaftler, Künstler oder Politiker – sind demzufolge immer an verschiedenen Orten gleichzeitig vorhanden. Der Zufall entscheidet dann leider meist darüber, an welcher Stelle der Nachlass zukünftig verwahrt und wissenschaftlich genutzt wird. Es stellt sich die Frage, ob wir auf solche Zufälle hoffen und warten sollen, oder ob wir nicht eher eine engagierte – gemeinsam zwischen den Archiven abgestimmte - Erwerbspolitik betrieben sollten. ------------ Beiträge zum Thema "Nachlässe an Universitäts- und Hochschularchiven sowie Archiven wissenschaftlicher Institutionen" im Rahmen der Frühjahrstagung der Fachgruppe 8: "Archivare an Hochschularchiven und Archiven wissenschaftlicher Institutionen" am 16./17. Juni an der Universität Potsdam.
Der Vortrag skizziert die Geschichte der deutschen Romanistik, gibt einen kurzen Überblick über den Stand 2003 im Bereich der Französischen Philologie und resümiert die fachlichen Herausforderungen im deutsch-französischen sowohl kulturellen wie politischen Kontext. Anschließend folgen drei Vorschläge zur Veränderung der Schulausbildung und Universitätslehre auf der Basis eines breiten Kulturverständnisses: 1. Einführung eines neuen allgemeinverbindlichen Schulfaches "Europa-Kunde" ("Connaissances de l'Europe"), das europaweit in Ergänzung oder Kooperation mit dem Fach Geschichte gelehrt werden sollte, 2. die systematische Ergänzung der traditionellen Literatur- und Sprachwissenschaft durch Kurse zu den deutsch-französischen Kulturbeziehungen, 3. die Ergänzung des traditionellen romanistischen Lehrkanons durch Seminare aus dem Bereich des Kulturmanagements.