Refine
Year of publication
- 2013 (311) (remove)
Document Type
- Doctoral Thesis (311) (remove)
Language
- English (174)
- German (135)
- French (1)
- Multiple languages (1)
Keywords
- Kinder (3)
- children (3)
- climate change (3)
- remote sensing (3)
- Adipositas (2)
- Arctic (2)
- Design Thinking (2)
- Eltern (2)
- Escherichia coli (2)
- Fernerkundung (2)
Institute
- Institut für Biochemie und Biologie (48)
- Institut für Geowissenschaften (35)
- Institut für Chemie (33)
- Institut für Physik und Astronomie (33)
- Wirtschaftswissenschaften (22)
- Institut für Ernährungswissenschaft (20)
- Hasso-Plattner-Institut für Digital Engineering gGmbH (15)
- Institut für Informatik und Computational Science (15)
- Department Linguistik (10)
- Department Psychologie (10)
- Öffentliches Recht (10)
- Sozialwissenschaften (9)
- Philosophische Fakultät (8)
- Institut für Umweltwissenschaften und Geographie (6)
- Historisches Institut (5)
- Institut für Mathematik (5)
- Institut für Romanistik (5)
- Bürgerliches Recht (4)
- Department Erziehungswissenschaft (4)
- Department Sport- und Gesundheitswissenschaften (4)
- Extern (4)
- Institut für Germanistik (3)
- Institut für Anglistik und Amerikanistik (2)
- Institut für Philosophie (2)
- Department Grundschulpädagogik (1)
- Institut für Jüdische Studien und Religionswissenschaft (1)
- Institut für Slavistik (1)
- Lehreinheit für Wirtschafts-Arbeit-Technik (1)
- Potsdam Transfer - Zentrum für Gründung, Innovation, Wissens- und Technologietransfer (1)
- Strafrecht (1)
Intensive Forschung hat in den vergangenen Jahrzehnten zu einer sehr detaillierten Charakterisierung des Geschmackssystems der Säugetiere geführt. Dennoch sind mit den bislang eingesetzten Methoden wichtige Fragestellungen unbeantwortet geblieben. Eine dieser Fragen gilt der Unterscheidung von Bitterstoffen. Die Zahl der Substanzen, die für den Menschen bitter schmecken und in Tieren angeborenes Aversionsverhalten auslösen, geht in die Tausende. Diese Substanzen sind sowohl von der chemischen Struktur als auch von ihrer Wirkung auf den Organismus sehr verschieden. Während viele Bitterstoffe potente Gifte darstellen, sind andere in den Mengen, die mit der Nahrung aufgenommen werden, harmlos oder haben sogar positive Effekte auf den Körper. Zwischen diesen Gruppen unterscheiden zu können, wäre für ein Tier von Vorteil. Ein solcher Mechanismus ist jedoch bei Säugetieren nicht bekannt. Das Ziel dieser Arbeit war die Untersuchung der Verarbeitung von Geschmacksinformation in der ersten Station der Geschmacksbahn im Mausgehirn, dem Nucleus tractus solitarii (NTS), mit besonderem Augenmerk auf der Frage nach der Diskriminierung verschiedener Bitterstoffe. Zu diesem Zweck wurde eine neue Untersuchungsmethode für das Geschmackssystem etabliert, die die Nachteile bereits verfügbarer Methoden umgeht und ihre Vorteile kombiniert. Die Arc-catFISH-Methode (cellular compartment analysis of temporal activity by fluorescent in situ hybridization), die die Charakterisierung der Antwort großer Neuronengruppen auf zwei Stimuli erlaubt, wurde zur Untersuchung geschmacksverarbeitender Zellen im NTS angewandt. Im Zuge dieses Projekts wurde erstmals eine stimulusinduzierte Arc-Expression im NTS gezeigt. Die ersten Ergebnisse offenbarten, dass die Arc-Expression im NTS spezifisch nach Stimulation mit Bitterstoffen auftritt und sich die Arc exprimierenden Neurone vornehmlich im gustatorischen Teil des NTS befinden. Dies weist darauf hin, dass Arc-Expression ein Marker für bitterverarbeitende gustatorische Neurone im NTS ist. Nach zweimaliger Stimulation mit Bittersubstanzen konnten überlappende, aber verschiedene Populationen von Neuronen beobachtet werden, die unterschiedlich auf die drei verwendeten Bittersubstanzen Cycloheximid, Chininhydrochlorid und Cucurbitacin I reagierten. Diese Neurone sind vermutlich an der Steuerung von Abwehrreflexen beteiligt und könnten so die Grundlage für divergentes Verhalten gegenüber verschiedenen Bitterstoffen bilden.
Automobildesigner haben als Gestaltungsexperten die Aufgabe, die Identität und damit die Werte einer Marke in Formen zu übersetzen, welche eine Vielzahl von Kunden ansprechen (Giannini & Monti, 2003; Karjalainen, 2002). Für diesen Übersetzungsprozess ist es zielführend, ästhetische Kundenbedürfnisse zu kennen, denn die Qualität einer Designlösung hängt auch davon ab, inwieweit der Designer Kundenbe-dürfnisse und damit das Designproblem richtig erfasst hat (Ulrich, 2006). Eine Grundlage hierfür entsteht durch eine erfolgreiche Designer-Nutzer-Interaktion und den Aufbau eines gemeinsamen Kontextwissens (Lee, Popovich, Blackler & Lee, 2009). Zwischen Designern und Kunden findet jedoch häufig kein direkter Austausch statt (Zeisel, 2006). Zudem belegen Befunde der Kunst- und Produktästhetikforschung, dass der Erwerb von gestalterischem Wissen und damit die Entwicklung ästhetischer Expertise mit Veränderungen der kognitiven Verarbeitung ästhetischer Objekte einhergeht, die sich in Wahrnehmung, Bewertung und Verhalten manifestieren. Damit ist auch zu erwarten, dass die Präferenzurteile von Designern und Kunden bei der ästhetischen Bewertung von Design nicht immer konvergieren. Ziel der vorliegenden Arbeit war daher die systematische Untersuchung dieser expertisebedingten Wahrnehmungs- und Bewertungsunterschiede zwischen designge-schulten und ungeschulten Personen bei der Betrachtung von Automobildesign. Damit sollten Perzeption, Verarbeitung und Bewertung von Automobildesign durch design-ungeschulte Personen transparenter gemacht und mit der Verarbeitung designgeschul-ter Personen verglichen werden, um einen Beitrag zur gemeinsamen Wissensbasis und damit einer erfolgreichen Designer-Nutzer-Interaktion zu leisten. Die theoretische Einbettung der Arbeit basierte auf dem Modell ästhetischer Erfahrung und ästheti-schen Urteilens von Leder, Belke, Oeberst und Augustin (2004), welches konkrete Annahmen zu Verarbeitungsunterschieden von ästhetischen Objekten zwischen Experten und Laien bietet, die bisher allerdings noch nicht umfassend geprüft wurden. Den ersten Schwerpunkt dieser Arbeit bildete die Untersuchung von Unter-schieden zwischen Designern und designungeschulten Rezipienten bei der Beschrei-bung und Bewertung auf dem Markt vorhandenen Fahrzeugdesigns. Dabei sollte auch geprüft werden, ob eine lexikalische Verbindung zwischen Beschreibungsattributen von Fahrzeugrezipienten und den postulierten Markenwerten von Automobilmarken hergestellt werden kann. Diesem ersten Untersuchungsanliegen wurde in zwei Studien nachgegangen: Studie I diente der Erhebung von Beschreibungsattributen mittels Triadenvergleich in Anlehnung an Kelly (1955). Es wurde geprüft, ob designgeschulte Teilnehmer produkti-ver verbalisieren, dabei anteilig mehr symbolbezogene als formbezogene Attribute generieren und innerhalb ihrer Gruppe häufiger gleiche Attribute nutzen als designun-geschulte Teilnehmer. Hierfür beschrieben 20 designgeschulte Probanden und 20 designungeschulte Probanden mit selbst gewählten Adjektiven die Unterschiede zwischen vier präsentierten Fahrzeugen. Die Gruppen nutzten dabei entgegen der Annahmen sehr ähnliche Attribute und unterschieden sich somit auch nicht in ihrer Verwendung symbolbezogener und formbezogener Attribute. Die generierten Attribute wurden mittels Prototypenansatz (Amelang & Zielinski, 2002) den ermittelten und nachfolgend kategorisierten Markenwerten von 10 Automobilherstellern zugeordnet, so dass sechs Skalen zur Erfassung der ästhetischen Wirkung von Fahrzeugen entstanden. In Studie II wurde ein diese sechs Skalen umfassender Fragebogen an einer Stichprobe von 83 Designern und Designstudierenden sowie 98 Probanden ohne Designausbildung in einer Onlinebefragung hinsichtlich Skalenkonsistenz geprüft. Außerdem wurden erste Annahmen aus dem Modell von Leder et al. (2004) abgeleitet und durch einen Vergleich der beiden Teilnehmergruppen hinsichtlich der Bewertung der vier präsentierten Fahrzeugmodelle für die Skalen mit guter interner Konsistenz (Attraktivität, Dynamik, Fortschritt, Qualität), sowie eines ästhetischen Gesamturteils, der benötigten Bewertungszeit und der Automobilaffinität überprüft. Hierbei vergaben Designstudierende und insbesondere ausgebildete Designer radikalere Bewertungen als Designlaien, benötigten mehr Zeit bei der Bewertung und waren automobilaffiner als die ungeschulten Befragungsteilnehmer. Den zweiten Schwerpunkt der Arbeit bildete eine konzeptionelle Zusammen-führung der Annahmen des Modells von Leder et al. (2004) und der Postulate zur Wirkung von Objekteigenschaften auf ästhetische Urteile (Berlyne, 1971; Martindale, 1988; Silvia, 2005b). Konkret sollte geprüft werden, welchen Einfluss marktrelevante Objekteigenschaften, wie z.B. das Ausmaß an Innovativität, auf die durch Expertise moderierte Bewertung von Design haben. In den Studien III und IV wurden hierfür systematisch bezüglich Innovativität und Balance gestufte Linienmodelle von Fahrzeu-gen präsentiert. In Studie III wurden die Modelle in einer Onlinebefragung durch 18 Designstudierende und 20 Studenten der Fahrzeugtechnik hinsichtlich Attraktivität, Innovativität und Balance bewertet. Im Einklang mit den Annahmen konnte gezeigt werden, dass sehr neuartiges Design von den designungeschulten Probanden als weniger attraktiv bewertet wird als von Betrachtern eines Designstudienganges. In Studie IV wurden neben den Ästhetikbewertungen zusätzlich das Blickverhal-ten und der affektiver Zustand der Versuchsteilnehmer in einem Messwiederholungs-design mit einer zwischengelagerten Phase elaborierter Designbewertung, in welcher der in Studie II geprüfte Fragebogen eingesetzt wurde, erhoben. An der Laborstudie nahmen je 11 Designer, Ingenieure, und Geisteswissenschaftler teil. Wiederum wurde innovatives Design von den designungeschulten Gruppen als weniger attraktiv bewertet. Dieser Unterschied reduzierte sich jedoch nach wiederholter Bewertung der Modelle. Die Manifestation expertisebedingten Blickverhaltens konnte nicht beobach-tet werden, wie auch die durch eine angenommene bessere Bewältigung einherge-hende positivere Stimmung oder höhere Zufriedenheit in der Expertengruppe. Gemeinsam mit den Befunden aus den Studien II und III wurde deutlich, dass Designausbildung und, noch ausgeprägter, Designexpertise neben einer höheren Attraktivitätsbewertung innovativen Designs auch zu einer differenzierteren Beurtei-lung von Innovativität führt. Dies wurde mit der Erweiterung des mentalen Schemas für Fahrzeuge durch die Beschäftigung mit vielfältigen Modellvarianten bereits während des Studiums interpretiert. Es wurden Hinweise auf eine stilbezogene, elaboriertere Verarbeitung von Fahrzeugdesign durch designgeschulte Betrachter beobachtet sowie eine mit Expertise einhergehende Autonomität ästhetischer Urteile als Ausdruck einer hohen ästhetischen Entwicklungsstufe (Parsons, 1987). Mit diesen bei unterschiedlichen Stichproben beobachteten, stabilen expertisebedingten Bewer-tungsunterschieden wurde eine begründete Basis für die geforderte Sensibilisierung für ästhetische Kundenbedürfnisse im Gestaltungsprozess geschaffen. Der in dieser Arbeit entwickelte Fragebogen kann hierbei für eine elaborierte Messung von Fahrzeugdesignpräferenzen, zum Vergleich der ästhetischen Wirkung mit den intendierten Markenwerten sowie für die Diskussion von Nutzereindrücken eingesetzt werden. Die Ergebnisse der vorliegenden Arbeiten tragen somit zur Erweiterung und Präzisierung des theoretischen Verständnisses von Ästhetikbewertungen bei und lassen sich gleichzeitig in die Praxis der Designausbildung und des Designprozesses übertragen.
Der W-Fragen-Erwerb stellt einen Teilbereich der kindlichen Syntaxentwicklung dar, die sich maßgeblich innerhalb der ersten drei Lebensjahre eines Kindes vollzieht. Eine wesentliche Rolle spielen dabei zwei Bewegungsoperationen, die sich auf die Position des Interrogativpronomens an die erste Stelle der W-Frage sowie die Position des Verbs an die zweite Stelle beziehen. In drei Studien wurde einerseits untersucht, ob deutschsprachige Kinder, die noch keine W-Fragen produzieren können, in der Lage sind, grammatische von ungrammatischen W-Fragen zu unterscheiden und andererseits, welche Leistungen sprachunauffällige und sprachauffällige deutschsprachige Kinder beim Verstehen und Korrigieren unterschiedlich komplexer W-Fragen (positive und negative W-Fragen) zeigen. Die Ergebnisse deuten auf ein frühes syntaktisches Wissen über W-Fragen im Spracherwerb hin und stützen damit die Annahme einer Kontinuität der kindlichen Grammatik zur Standardsprache. Auch scheinen sprachauffällige Kinder sich beim Erwerb von W-Fragen nicht qualitativ von sprachgesunden Kindern zu unterscheiden, sondern W-Fragen lediglich später korrekt umzusetzen. In beiden Populationen konnte ein syntaktischer Ökonomieeffekt beobachtet werden, der für eine spätere Umsetzung der Verbbewegung im Vergleich zur Bewegung des W-Elementes spricht.
Folgt tatsächlich aus einem liberalen Wertekanon eine generative Selbstbestimmung, eine weitgehende elterliche Handlungsfreiheit bei eugenischen Maßnahmen, wie es Vertreter einer „liberalen Eugenik“ versichern? Diese Arbeit diskutiert die Rolle Staates und die Handlungsspielräume der Eltern bei der genetischen Gestaltung von Nachkommen im Rahmen eines liberalen Wertverständnisses.
Den Schwerpunkt/Fokus der Betrachtungen liegt hier Maßnahmen des genetic enhancement.
Darüber hinaus wird auch das Verhältnis der „liberalen Eugenik“ zur „autoritären Eugenik“ neu beleuchtet.
Die Untersuchung beginnt bei der Analyse zentraler liberaler Werte und Normen, wie Freiheit, Autonomie und Gerechtigkeit und deren Funktionen in der „liberalen Eugenik“. Wobei nur sehr eingeschränkt von der „liberalen Eugenik“ gesprochen werden kann, sondern viel mehr von Varianten einer „liberalen Eugenik“.
Darüber hinaus wird in dieser Arbeit die historische Entwicklung der „liberalen“ und der „autoritären Eugenik“, speziell des Sozialdarwinismus, untersucht und verglichen, insbesondere im Hinblick auf liberale Werte und Normen und der generativen Selbstbestimmung.
Den Kern der Arbeit bildet der Vergleich der „liberalen Eugenik“ mit der „liberalen Erziehung“. Da hier die grundlegenden Aufgaben der Eltern, aber auch des Staates, analysiert und deren Verhältnis diskutiert wird.
Es zeigt sich, dass sich aus einem liberalen Wertverständnisses heraus keine umfangreiche generative Selbstbestimmung ableiten lässt, sondern sich viel mehr staatlich kontrollierte enge Grenzen bei eugenischen Maßnahmen zum Wohle der zukünftigen Person, begründen.
Zudem wurde der Weg zur autoritären Eugenik nicht durch die Abkehr von der generativen Selbstbestimmung geebnet, sondern viel mehr durch die Übertragung des Fortschrittsgedankens auf den Menschen selbst. Damit verliert die generative Selbstbestimmung auch ihre Funktion als Brandmauer gegen eine autoritäre Eugenik. Nicht der Verlust der generativen Selbstbestimmung, sondern viel mehr die Idee der Perfektionierung des Menschen muss kritisch betrachtet und letztlich abgelehnt werden.
Ohne generative Selbstbestimmung und einer Perfektionierung des Menschen, bleibt nur eine Basis-Eugenik, bei der die Entwicklungsfähigkeit des Menschen sichergestellt wird, nicht jedoch seine Verbesserung.
Darüber hinaus muss auch über eine Entwicklungsmöglichkeit des zukünftigen Menschen gesprochen werden, d. h. ein minimales Potential zu gesellschaftlicher Integration muss gegeben sein. Nur wenn tatsächlich keine Möglichkeiten seitens der Gesellschaft bestehen eine Person zu integrieren und dieser eine Entwicklungsmöglichkeit zu bieten, wären eugenische Maßnahmen als letztes Mittel akzeptabel.
In sedimentary basins, rock thermal conductivity can vary both laterally and vertically, thus altering the basin’s thermal structure locally and regionally. Knowledge of the thermal conductivity of geological formations and its spatial variations is essential, not only for quantifying basin evolution and hydrocarbon maturation processes, but also for understanding geothermal conditions in a geological setting. In conjunction with the temperature gradient, thermal conductivity represents the basic input parameter for the determination of the heat-flow density; which, in turn, is applied as a major input parameter in thermal modeling at different scales. Drill-core samples, which are necessary to determine thermal properties by laboratory measurements, are rarely available and often limited to previously explored reservoir formations. Thus, thermal conductivities of Mesozoic rocks in the North German Basin (NGB) are largely unknown. In contrast, geophysical borehole measurements are often available for the entire drilled sequence. Therefore, prediction equations to determine thermal conductivity based on well-log data are desirable. In this study rock thermal conductivity was investigated on different scales by (1) providing thermal-conductivity measurements on Mesozoic rocks, (2) evaluating and improving commonly applied mixing models which were used to estimate matrix and pore-filled rock thermal conductivities, and (3) developing new well-log based equations to predict thermal conductivity in boreholes without core control. Laboratory measurements are performed on sedimentary rock of major geothermal reservoirs in the Northeast German Basin (NEGB) (Aalenian, Rhaethian-Liassic, Stuttgart Fm., and Middle Buntsandstein). Samples are obtained from eight deep geothermal wells that approach depths of up to 2,500 m. Bulk thermal conductivities of Mesozoic sandstones range between 2.1 and 3.9 W/(m∙K), while matrix thermal conductivity ranges between 3.4 and 7.4 W/(m∙K). Local heat flow for the Stralsund location averages 76 mW/m², which is in good agreement to values reported previously for the NEGB. For the first time, in-situ bulk thermal conductivity is indirectly calculated for entire borehole profiles in the NEGB using the determined surface heat flow and measured temperature data. Average bulk thermal conductivity, derived for geological formations within the Mesozoic section, ranges between 1.5 and 3.1 W/(m∙K). The measurement of both dry- and water-saturated thermal conductivities allow further evaluation of different two-component mixing models which are often applied in geothermal calculations (e.g., arithmetic mean, geometric mean, harmonic mean, Hashin-Shtrikman mean, and effective-medium theory mean). It is found that the geometric-mean model shows the best correlation between calculated and measured bulk thermal conductivity. However, by applying new model-dependent correction, equations the quality of fit could be significantly improved and the error diffusion of each model reduced. The ‘corrected’ geometric mean provides the most satisfying results and constitutes a universally applicable model for sedimentary rocks. Furthermore, lithotype-specific and model-independent conversion equations are developed permitting a calculation of water-saturated thermal conductivity from dry-measured thermal conductivity and porosity within an error range of 5 to 10%. The limited availability of core samples and the expensive core-based laboratory measurements make it worthwhile to use petrophysical well logs to determine thermal conductivity for sedimentary rocks. The approach followed in this study is based on the detailed analyses of the relationships between thermal conductivity of rock-forming minerals, which are most abundant in sedimentary rocks, and the properties measured by standard logging tools. By using multivariate statistics separately for clastic, carbonate and evaporite rocks, the findings from these analyses allow the development of prediction equations from large artificial data sets that predict matrix thermal conductivity within an error of 4 to 11%. These equations are validated successfully on a comprehensive subsurface data set from the NGB. In comparison to the application of earlier published approaches formation-dependent developed for certain areas, the new developed equations show a significant error reduction of up to 50%. These results are used to infer rock thermal conductivity for entire borehole profiles. By inversion of corrected in-situ thermal-conductivity profiles, temperature profiles are calculated and compared to measured high-precision temperature logs. The resulting uncertainty in temperature prediction averages < 5%, which reveals the excellent temperature prediction capabilities using the presented approach. In conclusion, data and methods are provided to achieve a much more detailed parameterization of thermal models.
Vorstandsvergütung : eine rechtsökonomische Analyse zur Angemessenheit der Vorstandsvergütung
(2013)
Die kumulative Dissertation zur Projektdidaktik trägt den Titel „Von der Konzeption zur Praxis: Zur Entwicklung der Projektdidaktik am Oberstufen-Kolleg Bielefeld und ihre Impulsgebung und Modellbildung für das deutsche Regelschulwesen“. Die Dissertation versteht sich als beispielgebende Umsetzung und Implementierung der Projektdidaktik für das Regelschulsystem. Auf der Basis von 22 bereits erschienenen Publikationen und einer Monographie werden mit fünf methodischen Zugriffen (bildungshistorisch, dichte Beschreibung, Aktionsforschung, empirische Untersuchung an Regelschulen und Implementierungsforschung, s. Kapitel 1) in sieben Kapiteln (2- 8) des systematischen ersten Teils die Entwicklung der Unterrichtsform Projektunterricht in der BRD, Projektbegriff und Weiterentwicklung des Konzepts, Methodik, Bewertung sowie Organisation des Projektunterrichts am Oberstufen-Kolleg, der Versuchsschule des Landes NRW, in Auseinandersetzung mit der allgemeinen Projektdidaktik dargestellt sowie Formen und Verfahren der erprobten Implementierung in das Regelschulsystem präsentiert.
Ein Schlusskapitel (9) fasst die Ergebnisse zusammen. Im umfangreichen Anhang finden sich verschiedene Publikationen zu Aspekten der Projektdidaktik, auf die der systematische Teil jeweils Bezug nimmt.
Die bildungshistorische Analyse (Kapitel 2) untersucht das Verhältnis von pädagogischer Theorie und schulischer Praxis, die weder in Literatur und noch in Praxis genügend verbunden sind. Nach der Rezeption der gut erforschten Konzeptgeschichte pädagogischer Theorie in Anlehnung an Dewey und Kilpatrick wird durch eine erste Analyse der „Praxisgeschichte“ des Projektunterrichts auf ein Forschungsdesiderat hingewiesen, dies auch um die Projektpraxis am Oberstufen-Kolleg in Beziehung zu der in den Regelschulen setzen zu können. Dabei wurden seit 1975 sechs Entwicklungslinien herausgearbeitet: Start, Krise und ihre Überwindung durch Öffnung und Vernetzung (1975-1990), didaktisch-methodische Differenzierung und Notwendigkeit von Professionalisierung (ab 1990) sowie Schulentwicklung und Institutionalisierung (seit Ende der 1990er Jahre).
Projektunterricht besteht am Oberstufen-Kolleg seit der Gründung 1974 als fest eingerichtete Unterrichtsform (seit 2002 zweimal jährlich 2 Wochen) mit dem Ziel, für das Regelschulsystem die Projektdidaktik zu erproben und weiterzuentwickeln. Als wichtige praxisorientierte Ziele wurden ein praxistauglicher Begriff, Bildungswert und Kompetenzen im Unterschied zum Lehrgang herausgearbeitet (z.B. handlungs- und anwendungsorientierte Kompetenzen) und das Verhältnis zum Fachunterricht bestimmt (Kapitel 3). Letzteres wurde am Beispiel des Fachs Geschichte entwickelt und exemplarisch in Formen der Verzahnung dargestellt (Kapitel 6).
Auch für die methodische Dimension galt, die allgemeine Projektdidaktik weiterzuentwickeln durch ihre Abgrenzung zu anderen Methoden der Öffnung von Schule und Unterricht (Kapitel 4). Dabei wurde als zentrales methodisches Prinzip die Handlungsorientierung bestimmt sowie sieben Phasen und jeweilige Handlungsschritte festgelegt. Besonders Planung und Rollenwechsel bedürfen dabei besonderer Beachtung, um Selbsttätigkeit der ProjektteilnehmerInnen zu erreichen. Verschiedene methodische „Etüden“ ( z.B. Gruppenarbeit, recherchieren, sich öffentlich verhalten), handlungsorientierte Vorformen und projektorientiertes Arbeiten sollten die Vollform Projektunterricht vorbereiten helfen.
Die Bewertung von Projekten (Kapitel 5) stellt andere Anforderungen als der Lehrgang, weil sie unterschiedliche Bewertungsebenen (z.B. Prozessbedeutung, Produktbeurteilung, Gruppenbewertung) umfasst. Dazu sind am Oberstufen-Kolleg andere Bewertungsformen als die Ziffernnote entwickelt worden: z.B. ein „Reflexionsbericht“ als individuelle Rückmeldung von SchülerInnen und LehrerInnen und ein „Zertifikat“ für besondere Leistungen im Projekt.
Zentral für die Entwicklung von Projektunterricht ist jedoch die Organisationsfrage (Kapitel 7). Dazu bedarf es einer Organisationsgruppe Projekt, die die Unterrichtsform didaktisch betreut und in einem Hearing die angemeldeten Projekte berät. Das Oberstufen-Kolleg hat damit eine entwickelte „Projektkultur“ organisatorisch umgesetzt. Für eine empirische Untersuchung an sechs Regelschulen in Ostwestfalen ist dann eine idealtypische Merkmalsliste von schulischer „Projektkultur“ als Untersuchungsinstrument entstanden, das zugleich als Leitlinie für Schulentwicklung im Bereich Projektlernen in den Regelschulen dienen kann. Zu dieser Implementierung (Kapitel 8) wurden Konzepte und Erfahrungen vom Oberstufen-Kolleg für schulinterne und schulexterne Fortbildungsformen sowie eine exemplarische Fortbildungseinheit entwickelt. So konnten in zahlreichen Lehrerfortbildungen durch die Versuchsschule Impulse für das Regelschulsystem gegeben werden.
Requirements engineers have to elicit, document, and validate how stakeholders act and interact to achieve their common goals in collaborative scenarios. Only after gathering all information concerning who interacts with whom to do what and why, can a software system be designed and realized which supports the stakeholders to do their work. To capture and structure requirements of different (groups of) stakeholders, scenario-based approaches have been widely used and investigated. Still, the elicitation and validation of requirements covering collaborative scenarios remains complicated, since the required information is highly intertwined, fragmented, and distributed over several stakeholders. Hence, it can only be elicited and validated collaboratively. In times of globally distributed companies, scheduling and conducting workshops with groups of stakeholders is usually not feasible due to budget and time constraints. Talking to individual stakeholders, on the other hand, is feasible but leads to fragmented and incomplete stakeholder scenarios. Going back and forth between different individual stakeholders to resolve this fragmentation and explore uncovered alternatives is an error-prone, time-consuming, and expensive task for the requirements engineers. While formal modeling methods can be employed to automatically check and ensure consistency of stakeholder scenarios, such methods introduce additional overhead since their formal notations have to be explained in each interaction between stakeholders and requirements engineers. Tangible prototypes as they are used in other disciplines such as design, on the other hand, allow designers to feasibly validate and iterate concepts and requirements with stakeholders. This thesis proposes a model-based approach for prototyping formal behavioral specifications of stakeholders who are involved in collaborative scenarios. By simulating and animating such specifications in a remote domain-specific visualization, stakeholders can experience and validate the scenarios captured so far, i.e., how other stakeholders act and react. This interactive scenario simulation is referred to as a model-based virtual prototype. Moreover, through observing how stakeholders interact with a virtual prototype of their collaborative scenarios, formal behavioral specifications can be automatically derived which complete the otherwise fragmented scenarios. This, in turn, enables requirements engineers to elicit and validate collaborative scenarios in individual stakeholder sessions – decoupled, since stakeholders can participate remotely and are not forced to be available for a joint session at the same time. This thesis discusses and evaluates the feasibility, understandability, and modifiability of model-based virtual prototypes. Similarly to how physical prototypes are perceived, the presented approach brings behavioral models closer to being tangible for stakeholders and, moreover, combines the advantages of joint stakeholder sessions and decoupled sessions.
Escherichia (E.) coli ist als kommensales Bakterium ein wichtiger Bestandteil des Mikrobioms von Säugern, jedoch zudem der häufigste Infektionserreger des Menschen. Entsprechend des Infektionsortes werden intestinal (InPEC) und extraintestinal pathogene E. coli (ExPEC) unterschieden. Die Pathogenese von E. coli-Infektionen ist durch Virulenzfaktoren determiniert, welche von jeweils spezifischen virulenzassoziierten Genen (inVAGs und exVAGs) kodiert werden. Häufig werden exVAGs auch in E. coli-Isolaten aus dem Darm gesunder Wirte nachgewiesen. Dies führte zu der Vermutung, dass exVAGs die intestinale Kolonisierung des Wirtes durch E. coli unterstützen. Das Hauptziel dieser Arbeit bestand darin, das Wissen über den Einfluss von exVAGs auf die Besiedlung und damit die Adhäsion von E. coli an Epithelzellen des Darmtraktes zu erweitern. Die Durchführung einer solch umfassenden E. coli-Populationsstudie erforderte die Etablierung neuer Screeningmethoden. Für die genotypische Charakterisierung wurden mikropartikelbasierte Multiplex-PCR-Assays zum Nachweis von 44 VAGs und der Phylogenie etabliert. Für die phänotypische Charakterisierung wurden Adhäsions- und Zytotoxizitätsassays etabliert. Die Screeningmethoden basieren auf der VideoScan-Technologie, einem automatisierten bildbasierten Multifluoreszenzdetektionssystem. Es wurden 398 E. coli-Isolate aus 13 Wildsäugerarten und 5 Wildvogelarten sowie aus gesunden und harnwegserkrankten Menschen und Hausschweinen charakterisiert. Die Adhäsionsassays hatten zum Ziel, sowohl die Adhäsionsraten als auch die Adhäsionsmuster der 317 nicht hämolytischen Isolate auf 5 Epithelzelllinien zu bestimmen. Die Zytotoxizität der 81 hämolytischen Isolate wurde in Abhängigkeit der Inkubationszeit auf 4 Epithelzelllinien geprüft. In den E. coli-Isolaten wurde eine Reihe von VAGs nachgewiesen. Potentielle InPEC, insbesondere shigatoxinproduzierende und enteropathogene E. coli wurden aus Menschen, Hausschweinen und Wildtieren, vor allem aus Rehen und Feldhasen isoliert. exVAGs wurden mit stark variierender Prävalenz in Isolaten aus allen Arten detektiert. Die größte Anzahl und das breiteste Spektrum an exVAGs wurde in Isolaten aus Urin harnwegserkrankter Menschen, gefolgt von Isolaten aus Dachsen und Rehen nachgewiesen. In Isolaten der phylogenetischen Gruppe B2 wurden mehr exVAGs detektiert als in den Isolaten der phylogenetischen Gruppen A, B1 und D. Die Ergebnisse der Adhäsionsassays zeigten, dass die meisten Isolate zelllinien-, gewebe- oder wirtsspezifisch adhärierten. Ein Drittel der Isolate adhärierte an keiner Zelllinie und nur zwei Isolate adhärierten stark an allen Zelllinien. Grundsätzlich adhärierten mehr Isolate an humanen sowie an intestinalen Zelllinien. Besonders Isolate aus Eichhörnchen und Amseln sowie aus Urin harnwegserkrankter Menschen und Hausschweine waren in der Lage, stark zu adhärieren. Hierbei bildeten die Isolate als Adhäsionsmuster diffuse Adhäsion, Mikrokolonien, Ketten und Agglomerationen. Mittels statistischer Analysen wurden Assoziationen zwischen exVAGs und einer hohen Adhäsionsrate ersichtlich. So war beispielsweise das Vorkommen von afa/dra mit einer höheren Adhäsionsrate auf Caco-2- und 5637-Zellen und von sfa/foc auf IPEC-J2-Zellen assoziiert. Die Ergebnisse der Zytotoxizitätsassays zeigten eine sehr starke und zeitabhängige Zerstörung der Monolayer aller Epithelzelllinien durch die α-Hämolysin-positiven Isolate. Auffallend war die hohe Toxizität hämolytischer Isolate aus Wildtieren gegenüber den humanen Zelllinien. Mit den innerhalb dieser Arbeit entwickelten Screeningmethoden war es möglich, große Mengen an Bakterien zu charakterisieren. Es konnte ein Überblick über die Verbreitung von VAGs in E. coli aus unterschiedlichen Wirten gewonnen werden. Besonders Wildtiere wurden sowohl durch den Nachweis von VAGs in den entsprechenden Isolaten, verbunden mit deren Adhäsionsfähigkeit und ausgeprägter Zytotoxizität als Reservoire pathogener E. coli identifiziert. Ebenso wurde eine zelllinienspezifische Adhäsion von Isolaten mit bestimmten exVAGs deutlich. Damit konnte der mögliche Einfluss von exVAGs auf die intestinale Kolonisierung bestätigt werden. In weiterführenden Arbeiten sind jedoch Expressions- und Funktionsanalysen der entsprechenden Proteine unerlässlich. Es wird anhand der Mikrokoloniebildung durch kommensale E. coli vermutet, dass Adhäsionsmuster und demzufolge Kolonisierungsstrategien, die bisher pathogenen E. coli zugeschrieben wurden, eher als generelle Kolonisierungsstrategien zu betrachten sind. Das E. coli-α-Hämolysin wirkt im Allgemeinen zytotoxisch auf Epithelzellen. Ein in der Fachliteratur diskutierter adhäsionsunterstützender Mechanismus dieses Toxins ist demnach fragwürdig. Innerhalb dieser Arbeit konnte gezeigt werden, dass die entwickelten Screeningmethoden umfassende Analysen einer großen Anzahl an E. coli-Isolaten ermöglichen.
The main intention of the PhD project was to create a varve chronology for the Suigetsu Varves 2006' (SG06) composite profile from Lake Suigetsu (Japan) by thin section microscopy. The chronology was not only to provide an age-scale for the various palaeo-environmental proxies analysed within the SG06 project, but also and foremost to contribute, in combination with the SG06 14C chronology, to the international atmospheric radiocarbon calibration curve (IntCal). The SG06 14C data are based on terrestrial leaf fossils and therefore record atmospheric 14C values directly, avoiding the corrections necessary for the reservoir ages of the marine datasets, which are currently used beyond the tree-ring limit in the IntCal09 dataset (Reimer et al., 2009). The SG06 project is a follow up of the SG93 project (Kitagawa & van der Plicht, 2000), which aimed to produce an atmospheric calibration dataset, too, but suffered from incomplete core recovery and varve count uncertainties. For the SG06 project the complete Lake Suigetsu sediment sequence was recovered continuously, leaving the task to produce an improved varve count. Varve counting was carried out using a dual method approach utilizing thin section microscopy and micro X-Ray Fluorescence (µXRF). The latter was carried out by Dr. Michael Marshall in cooperation with the PhD candidate. The varve count covers 19 m of composite core, which corresponds to the time frame from ≈10 to ≈40 kyr BP. The count result showed that seasonal layers did not form in every year. Hence, the varve counts from either method were incomplete. This rather common problem in varve counting is usually solved by manual varve interpolation. But manual interpolation often suffers from subjectivity. Furthermore, sedimentation rate estimates (which are the basis for interpolation) are generally derived from neighbouring, well varved intervals. This assumes that the sedimentation rates in neighbouring intervals are identical to those in the incompletely varved section, which is not necessarily true. To overcome these problems a novel interpolation method was devised. It is computer based and automated (i.e. avoids subjectivity and ensures reproducibility) and derives the sedimentation rate estimate directly from the incompletely varved interval by statistically analysing distances between successive seasonal layers. Therefore, the interpolation approach is also suitable for sediments which do not contain well varved intervals. Another benefit of the novel method is that it provides objective interpolation error estimates. Interpolation results from the two counting methods were combined and the resulting chronology compared to the 14C chronology from Lake Suigetsu, calibrated with the tree-ring derived section of IntCal09 (which is considered accurate). The varve and 14C chronology showed a high degree of similarity, demonstrating that the novel interpolation method produces reliable results. In order to constrain the uncertainties of the varve chronology, especially the cumulative error estimates, U-Th dated speleothem data were used by linking the low frequency 14C signal of Lake Suigetsu and the speleothems, increasing the accuracy and precision of the Suigetsu calibration dataset. The resulting chronology also represents the age-scale for the various palaeo-environmental proxies analysed in the SG06 project. One proxy analysed within the PhD project was the distribution of event layers, which are often representatives of past floods or earthquakes. A detailed microfacies analysis revealed three different types of event layers, two of which are described here for the first time for the Suigetsu sediment. The types are: matrix supported layers produced as result of subaqueous slope failures, turbidites produced as result of landslides and turbidites produced as result of flood events. The former two are likely to have been triggered by earthquakes. The vast majority of event layers was related to floods (362 out of 369), which allowed the construction of a respective chronology for the last 40 kyr. Flood frequencies were highly variable, reaching their greatest values during the global sea level low-stand of the Glacial, their lowest values during Heinrich Event 1. Typhoons affecting the region represent the most likely control on the flood frequency, especially during the Glacial. However, also local, non-climatic controls are suggested by the data. In summary, the work presented here expands and revises knowledge on the Lake Suigetsu sediment and enabls the construction of a far more precise varve chronology. The 14C calibration dataset is the first such derived from lacustrine sediments to be included into the (next) IntCal dataset. References: Kitagawa & van der Plicht, 2000, Radiocarbon, Vol 42(3), 370-381 Reimer et al., 2009, Radiocarbon, Vol 51(4), 1111-1150
Large Central European flood events of the past have demonstrated that flooding can affect several river basins at the same time leading to catastrophic economic and humanitarian losses that can stretch emergency resources beyond planned levels of service. For Germany, the spatial coherence of flooding, the contributing processes and the role of trans-basin floods for a national risk assessment is largely unknown and analysis is limited by a lack of systematic data, information and knowledge on past events. This study investigates the frequency and intensity of trans-basin flood events in Germany. It evaluates the data and information basis on which knowledge about trans-basin floods can be generated in order to improve any future flood risk assessment. In particu-lar, the study assesses whether flood documentations and related reports can provide a valuable data source for understanding trans-basin floods. An adaptive algorithm was developed that systematically captures trans-basin floods using series of mean daily discharge at a large number of sites of even time series length (1952-2002). It identifies the simultaneous occurrence of flood peaks based on the exceedance of an initial threshold of a 10 year flood at one location and consecutively pools all causally related, spatially and temporally lagged peak recordings at the other locations. A weighted cumulative index was developed that accounts for the spatial extent and the individual flood magnitudes within an event and allows quantifying the overall event severity. The parameters of the method were tested in a sensitivity analysis. An intensive study on sources and ways of information dissemination of flood-relevant publications in Germany was conducted. Based on the method of systematic reviews a strategic search approach was developed to identify relevant documentations for each of the 40 strongest trans-basin flood events. A novel framework for assessing the quality of event specific flood reports from a user’s perspective was developed and validated by independent peers. The framework was designed to be generally applicable for any natural hazard type and assesses the quality of a document addressing accessibility as well as representational, contextual, and intrinsic dimensions of quality. The analysis of time-series of mean daily discharge resulted in the identification of 80 trans-basin flood events within the period 1952-2002 in Germany. The set is dominated by events that were recorded in the hydrological winter (64%); 36% occurred during the summer months. The occurrence of floods is characterised by a distinct clustering in time. Dividing the study period into two sub-periods, we find an increase in the percentage of winter events from 58% in the first to 70.5% in the second sub-period. Accordingly, we find a significant increase in the number of extreme trans-basin floods in the second sub-period. A large body of 186 flood relevant documentations was identified. For 87.5% of the 40 strongest trans-basin floods in Germany at least one report has been found and for the most severe floods a substantial amount of documentation could be obtained. 80% of the material can be considered grey literature (i.e. literature not controlled by commercial publishers). The results of the quality assessment show that the majority of flood event specific reports are of a good quality, i.e. they are well enough drafted, largely accurate and objective, and contain a substantial amount of information on the sources, pathways and receptors/consequences of the floods. The inclusion of this information in the process of knowledge building for flood risk assessment is recommended. Both the results as well as the data produced in this study are openly accessible and can be used for further research. The results of this study contribute to an improved spatial risk assessment in Germany. The identified set of trans-basin floods provides the basis for an assessment of the chance that flooding occurs simultaneously at a number of sites. The information obtained from flood event documentation can usefully supplement the analysis of the processes that govern flood risk.
User-centered design processes are the first choice when new interactive systems or services are developed to address real customer needs and provide a good user experience. Common tools for collecting user research data, conducting brainstormings, or sketching ideas are whiteboards and sticky notes. They are ubiquitously available, and no technical or domain knowledge is necessary to use them. However, traditional pen and paper tools fall short when saving the content and sharing it with others unable to be in the same location. They are also missing further digital advantages such as searching or sorting content. Although research on digital whiteboard and sticky note applications has been conducted for over 20 years, these tools are not widely adopted in company contexts. While many research prototypes exist, they have not been used for an extended period of time in a real-world context. The goal of this thesis is to investigate what the enablers and obstacles for the adoption of digital whiteboard systems are. As an instrument for different studies, we developed the Tele-Board software system for collaborative creative work. Based on interviews, observations, and findings from former research, we tried to transfer the analog way of working to the digital world. Being a software system, Tele-Board can be used with a variety of hardware and does not depend on special devices. This feature became one of the main factors for adoption on a larger scale. In this thesis, I will present three studies on the use of Tele-Board with different user groups and foci. I will use a combination of research methods (laboratory case studies and data from field research) with the overall goal of finding out when a digital whiteboard system is used and in which cases not. Not surprisingly, the system is used and accepted if a user sees a main benefit that neither analog tools nor other applications can offer. However, I found that these perceived benefits are very different for each user and usage context. If a tool provides possibilities to use in different ways and with different equipment, the chances of its adoption by a larger group increase. Tele-Board has now been in use for over 1.5 years in a global IT company in at least five countries with a constantly growing user base. Its use, advantages, and disadvantages will be described based on 42 interviews and usage statistics from server logs. Through these insights and findings from laboratory case studies, I will present a detailed analysis of digital whiteboard use in different contexts with design implications for future systems.
Within the course of this thesis, I have investigated the complex interplay between electron and lattice dynamics in nanostructures of perovskite oxides. Femtosecond hard X-ray pulses were utilized to probe the evolution of atomic rearrangement directly, which is driven by ultrafast optical excitation of electrons. The physics of complex materials with a large number of degrees of freedom can be interpreted once the exact fingerprint of ultrafast lattice dynamics in time-resolved X-ray diffraction experiments for a simple model system is well known. The motion of atoms in a crystal can be probed directly and in real-time by femtosecond pulses of hard X-ray radiation in a pump-probe scheme. In order to provide such ultrashort X-ray pulses, I have built up a laser-driven plasma X-ray source. The setup was extended by a stable goniometer, a two-dimensional X-ray detector and a cryogen-free cryostat. The data acquisition routines of the diffractometer for these ultrafast X-ray diffraction experiments were further improved in terms of signal-to-noise ratio and angular resolution. The implementation of a high-speed reciprocal-space mapping technique allowed for a two-dimensional structural analysis with femtosecond temporal resolution. I have studied the ultrafast lattice dynamics, namely the excitation and propagation of coherent phonons, in photoexcited thin films and superlattice structures of the metallic perovskite SrRuO3. Due to the quasi-instantaneous coupling of the lattice to the optically excited electrons in this material a spatially and temporally well-defined thermal stress profile is generated in SrRuO3. This enables understanding the effect of the resulting coherent lattice dynamics in time-resolved X-ray diffraction data in great detail, e.g. the appearance of a transient Bragg peak splitting in both thin films and superlattice structures of SrRuO3. In addition, a comprehensive simulation toolbox to calculate the ultrafast lattice dynamics and the resulting X-ray diffraction response in photoexcited one-dimensional crystalline structures was developed in this thesis work. With the powerful experimental and theoretical framework at hand, I have studied the excitation and propagation of coherent phonons in more complex material systems. In particular, I have revealed strongly localized charge carriers after above-bandgap femtosecond photoexcitation of the prototypical multiferroic BiFeO3, which are the origin of a quasi-instantaneous and spatially inhomogeneous stress that drives coherent phonons in a thin film of the multiferroic. In a structurally imperfect thin film of the ferroelectric Pb(Zr0.2Ti0.8)O3, the ultrafast reciprocal-space mapping technique was applied to follow a purely strain-induced change of mosaicity on a picosecond time scale. These results point to a strong coupling of in- and out-of-plane atomic motion exclusively mediated by structural defects.
The nutrient exchange between plant and fungus is the key element of the arbuscular mycorrhizal (AM) symbiosis. The fungus improves the plant’s uptake of mineral nutrients, mainly phosphate, and water, while the plant provides the fungus with photosynthetically assimilated carbohydrates. Still, the knowledge about the mechanisms of the nutrient exchange between the symbiotic partners is very limited. Therefore, transport processes of both, the plant and the fungal partner, are investigated in this study. In order to enhance the understanding of the molecular basis underlying this tight interaction between the roots of Medicago truncatula and the AM fungus Rhizophagus irregularis, genes involved in transport processes of both symbiotic partners are analysed here. The AM-specific regulation and cell-specific expression of potential transporter genes of M. truncatula that were found to be specifically regulated in arbuscule-containing cells and in non-arbusculated cells of mycorrhizal roots was confirmed. A model for the carbon allocation in mycorrhizal roots is suggested, in which carbohydrates are mobilized in non-arbusculated cells and symplastically provided to the arbuscule-containing cells. New insights into the mechanisms of the carbohydrate allocation were gained by the analysis of hexose/H+ symporter MtHxt1 which is regulated in distinct cells of mycorrhizal roots. Metabolite profiling of leaves and roots of a knock-out mutant, hxt1, showed that it indeed does have an impact on the carbohydrate balance in the course of the symbiosis throughout the whole plant, and on the interaction with the fungal partner. The primary metabolite profile of M. truncatula was shown to be altered significantly in response to mycorrhizal colonization. Additionally, molecular mechanisms determining the progress of the interaction in the fungal partner of the AM symbiosis were investigated. The R. irregularis transcriptome in planta and in extraradical tissues gave new insight into genes that are differentially expressed in these two fungal tissues. Over 3200 fungal transcripts with a significantly altered expression level in laser capture microdissection-collected arbuscules compared to extraradical tissues were identified. Among them, six previously unknown specifically regulated potential transporter genes were found. These are likely to play a role in the nutrient exchange between plant and fungus. While the substrates of three potential MFS transporters are as yet unknown, two potential sugar transporters are might play a role in the carbohydrate flow towards the fungal partner. In summary, this study provides new insights into transport processes between plant and fungus in the course of the AM symbiosis, analysing M. truncatula on the transcript and metabolite level, and provides a dataset of the R. irregularis transcriptome in planta, providing a high amount of new information for future works.
Thermodynamic stability of the a-Helical membrane-interacting protein mistic in detergent micelles
(2013)
Galaxy clusters are the largest known gravitationally bound objects, their study is important for both an intrinsic understanding of their systems and an investigation of the large scale structure of the universe. The multi- component nature of galaxy clusters offers multiple observable signals across the electromagnetic spectrum. At X-ray wavelengths, galaxy clusters are simply identified as X-ray luminous, spatially extended, and extragalactic sources. X-ray observations offer the most powerful technique for constructing cluster catalogues. The main advantages of the X-ray cluster surveys are their excellent purity and completeness and the X-ray observables are tightly correlated with mass, which is indeed the most fundamental parameter of clusters. In my thesis I have conducted the 2XMMi/SDSS galaxy cluster survey, which is a serendipitous search for galaxy clusters based on the X-ray extended sources in the XMM-Newton Serendipitous Source Catalogue (2XMMi-DR3). The main aims of the survey are to identify new X-ray galaxy clusters, investigate their X-ray scaling relations, identify distant cluster candidates, and study the correlation of the X-ray and optical properties. The survey is constrained to those extended sources that are in the footprint of the Sloan Digital Sky Survey (SDSS) in order to be able to identify the optical counterparts as well as to measure their redshifts that are mandatory to measure their physical properties. The overlap area be- tween the XMM-Newton fields and the SDSS-DR7 imaging, the latest SDSS data release at the starting of the survey, is 210 deg^2. The survey comprises 1180 X-ray cluster candidates with at least 80 background-subtracted photon counts, which passed the quality control process. To measure the optical redshifts of the X-ray cluster candidates, I used three procedures; (i) cross-matching these candidates with the recent and largest optically selected cluster catalogues in the literature, which yielded the photometric redshifts of about a quarter of the X-ray cluster candidates. (ii) I developed a finding algorithm to search for overdensities of galaxies at the positions of the X-ray cluster candidates in the photometric redshift space and to measure their redshifts from the SDSS-DR8 data, which provided the photometric redshifts of 530 groups/clusters. (iii) I developed an algorithm to identify the cluster candidates associated with spectroscopically targeted Luminous Red Galaxies (LRGs) in the SDSS-DR9 and to measure the cluster spectroscopic redshift, which provided 324 groups and clusters with spectroscopic confirmation based on spectroscopic redshift of at least one LRG. In total, the optically confirmed cluster sample comprises 574 groups and clusters with redshifts (0.03 ≤ z ≤ 0.77), which is the largest X-ray selected cluster catalogue to date based on observations from the current X-ray observatories (XMM-Newton, Chandra, Suzaku, and Swift/XRT). Among the cluster sample, about 75 percent are newly X-ray discovered groups/clusters and 40 percent are new systems to the literature. To determine the X-ray properties of the optically confirmed cluster sample, I reduced and analysed their X-ray data in an automated way following the standard pipelines of processing the XMM-Newton data. In this analysis, I extracted the cluster spectra from EPIC(PN, MOS1, MOS2) images within an optimal aperture chosen to maximise the signal-to-noise ratio. The spectral fitting procedure provided the X-ray temperatures kT (0.5 - 7.5 keV) for 345 systems that have good quality X-ray data. For all the optically confirmed cluster sample, I measured the physical properties L500 (0.5 x 10^42 – 1.2 x 10^45 erg s-1 ) and M500 (1.1 x 10^13 – 4.9 x 10^14 M⊙) from an iterative procedure using published scaling relations. The present X-ray detected groups and clusters are in the low and intermediate luminosity regimes apart from few luminous systems, thanks to the XMM-Newton sensitivity and the available XMM-Newton deep fields The optically confirmed cluster sample with measurements of redshift and X-ray properties can be used for various astrophysical applications. As a first application, I investigated the LX - T relation for the first time based on a large cluster sample of 345 systems with X-ray spectroscopic parameters drawn from a single survey. The current sample includes groups and clusters with wide ranges of redshifts, temperatures, and luminosities. The slope of the relation is consistent with the published ones of nearby clusters with higher temperatures and luminosities. The derived relation is still much steeper than that predicted by self-similar evolution. I also investigated the evolution of the slope and the scatter of the LX - T relation with the cluster redshift. After excluding the low luminosity groups, I found no significant changes of the slope and the intrinsic scatter of the relation with redshift when dividing the sample into three redshift bins. When including the low luminosity groups in the low redshift subsample, I found its LX - T relation becomes after than the relation of the intermediate and high redshift subsamples. As a second application of the optically confirmed cluster sample from our ongoing survey, I investigated the correlation between the cluster X-ray and the optical parameters that have been determined in a homogenous way. Firstly, I investigated the correlations between the BCG properties (absolute magnitude and optical luminosity) and the cluster global proper- ties (redshift and mass). Secondly, I computed the richness and the optical luminosity within R500 of a nearby subsample (z ≤ 0.42, with a complete membership detection from the SDSS data) with measured X-ray temperatures from our survey. The relation between the estimated optical luminosity and richness is also presented. Finally, the correlation between the cluster optical properties (richness and luminosity) and the cluster global properties (X-ray luminosity, temperature, mass) are investigated.
Pronoun resolution normally takes place without conscious effort or awareness, yet the processes behind it are far from straightforward. A large number of cues and constraints have previously been recognised as playing a role in the identification and integration of potential antecedents, yet there is considerable debate over how these operate within the resolution process. The aim of this thesis is to investigate how the parser handles multiple antecedents in order to understand more about how certain information sources play a role during pronoun resolution. I consider how both structural information and information provided by the prior discourse is used during online processing. This is investigated through several eye tracking during reading experiments that are complemented by a number of offline questionnaire experiments. I begin by considering how condition B of the Binding Theory (Chomsky 1981; 1986) has been captured in pronoun processing models; some researchers have claimed that processing is faithful to syntactic constraints from the beginning of the search (e.g. Nicol and Swinney 1989), while others have claimed that potential antecedents which are ruled out on structural grounds nonetheless affect processing, because the parser must also pay attention to a potential antecedent’s features (e.g. Badecker and Straub 2002). My experimental findings demonstrate that the parser is sensitive to the subtle changes in syntactic configuration which either allow or disallow pronoun reference to a local antecedent, and indicate that the parser is normally faithful to condition B at all stages of processing. Secondly, I test the Primitives of Binding hypothesis proposed by Koornneef (2008) based on work by Reuland (2001), which is a modular approach to pronoun resolution in which variable binding (a semantic relationship between pronoun and antecedent) takes place before coreference. I demonstrate that a variable-binding (VB) antecedent is not systematically considered earlier than a coreference (CR) antecedent online. I then go on to explore whether these findings could be attributed to the linear order of the antecedents, and uncover a robust recency preference both online and offline. I consider what role the factor of recency plays in pronoun resolution and how it can be reconciled with the first-mention advantage (Gernsbacher and Hargreaves 1988; Arnold 2001; Arnold et al., 2007). Finally, I investigate how aspects of the prior discourse affect pronoun resolution. Prior discourse status clearly had an effect on pronoun resolution, but an antecedent’s appearance in the previous context was not always facilitative; I propose that this is due to the number of topic switches that a reader must make, leading to a lack of discourse coherence which has a detrimental effect on pronoun resolution. The sensitivity of the parser to structural cues does not entail that cue types can be easily separated into distinct sequential stages, and I therefore propose that the parser is structurally sensitive but not modular. Aspects of pronoun resolution can be captured within a parallel constraints model of pronoun resolution, however, such a model should be sensitive to the activation of potential antecedents based on discourse factors, and structural cues should be strongly weighted.
Rhythm is a temporal and systematic organization of acoustic events in terms of prominence, timing and grouping, helping to structure our most basic experiences, such as body movement, music and speech. In speech, rhythm groups auditory events, e.g., sounds and pauses, together into words, making their boundaries acoustically prominent and aiding word segmentation and recognition by the hearer. After word recognition, the hearer is able to retrieve word meaning form his mental lexicon, integrating it with information from other linguistic domains, such as semantics, syntax and pragmatics, until comprehension is achieved. The importance of speech rhythm, however, is not restricted to word segmentation and recognition only. Beyond the word level rhythm continues to operate as an organization device, interacting with different linguistic domains, such as syntax and semantics, and grouping words into larger prosodic constituents, organized in a prosodic hierarchy. This dissertation investigates the function of speech rhythm as a sentence segmentation device during syntactic ambiguity processing, possible limitations on its use, i.e., in the context of second language processing, and its transferability as cognitive skill to the music domain.
The evolution of most orogens typically records cogenetic shortening and extension. Pervasive normal faulting in an orogen, however, has been related to late syn- and post-collisional stages of mountain building with shortening focused along the peripheral sectors of the orogen. While extensional processes constitute an integral part of orogenic evolution, the spatiotemporal characteristics and the kinematic linkage of structures related to shortening and extension in the core regions of the orogen are often not well known. Related to the India-Eurasia collision, the Himalaya forms the southern margin of the Tibetan Plateau and constitutes the most prominent Cenozoic type example of a collisional orogen. While thrusting is presently observed along the foothills of the orogen, several generations of extensional structures have been detected in the internal, high-elevation regions, both oriented either parallel or perpendicular to the strike of the orogen. In the NW Indian Himalaya, earthquake focal mechanisms, seismites and ubiquitous normal faulting in Quaternary deposits, and regional GPS measurements reveal ongoing E-W extension. In contrast to other extensional structures observed in the Himalaya, this extension direction is neither parallel nor perpendicular to the NE-SW regional shortening direction. In this study, I took advantage of this obliquity between the trend of the orogen and structures related to E-W oriented extension in order to address the question of the driving forces of different extension directions. Thus, extension might be triggered triggered by processes within the Tibetan Plateau or originates from the curvature of the Himalayan orogen. In order to elaborate on this topic, I present new fault-kinematic data based on systematic measurements of approximately 2000 outcrop-scale brittle fault planes with displacements of up to several centimeters that cover a large area of the NW Indian Himalaya. This new data set together with field observations relevant for relative chronology allows me to distinguish six different deformation styles. One of the main results are that the overall strain pattern derived from this data reflects the regionally important contractional deformation pattern very well, but also reveals significant extensional deformation. In total, I was able to identify six deformation styles, most of which are temporally and spatially linked and represent protracted shortening, but also significant extensional directions. For example, this is the first data set where a succession of both, arc-normal and E-W extension have been documented in the Himalaya. My observations also furnish the basis for a detailed overview of the younger extensional deformation history in the NW Indian Himalaya. Field and remote-sensing based geomorphic analyses, and geochronologic 40Ar/39Ar data on synkinematic muscovites along normal faults help elucidate widespread E-W extension in the NW Indian Himalaya which must have started at approximately 14-16 Ma, if not earlier. In addition, I documented and mapped fault scarps in Quaternary sedimentary deposits using satellite imagery and field inspection. Furthermore, I made field observations of regional normal faults, compiled structures from geological maps and put them in a regional context. Finally, I documented seismites in lake sediments close to the currently most active normal fault in the study area in order to extend the (paleo) seismic record of this particular fault. Taken together, this data sets document that E-W extension is the dominant active deformation style in the internal parts of the orogen. In addition, the combined field, geomorphic and remote-sensing data sets prove that E-W extension occurs in a much more larger region toward the south and west than the seismicity data have suggested. In conclusion, the data presented here reveal the importance of extension in a region, which is still dominated by ongoing collision and shortening. The regional fault distribution and cross-cutting relationships suggest that extension parallel and perpendicular to the strike of the orogen are an integral part of the southward propagation of the active thrust front and the associated lateral growth of the Himalayan arc. In the light of a wide range of models proposed for extension in the Himalaya and the Tibetan plateau, I propose that E-W extension in the NW Indian Himalaya is transferred from the Tibetan Plateau due the inability of the Karakorum fault (KF) to adequately accommodate ongoing E-W extension on the Tibetan Plateau. Furthermore, in line with other observations from Tibet, the onset of E-W normal faulting in the NW Himalaya may also reflect the attainment of high topography in this region, which generated crustal stresses conducive to spatially extensive extension.
Intra-continental mountain belts typically form as a result of tectonic forces associated with distant plate collisions. In general, each mountain belt has a distinctive morphology and orogenic evolution that is highly dependent on the unique distribution and geometries of inherited structures and other crustal weaknesses. In this thesis, I have investigated the complex and irregular Cenozoic orogenic evolution of the Central Kyrgyz Tien Shan in Central Asia, which is presently one of the most active intra-continental mountain belts in the world. This work involved combining a broad array of datasets, including thermochronologic, magnetostratigraphic, sediment provenance and stable isotope data, to identify and date various changes in tectonic deformation, climate and surface processes. Many of these changes are linked and can ultimately be related to regional-scale processes that altered the orogenic evolution of the Central Kyrgyz Tien Shan. The Central Kyrgyz Tien Shan contains a sub-parallel series of structures that were reactivated in the late Cenozoic in response to the tectonic forces associated with the distant India-Eurasia collision. Over time, slip on the various reactivated structures created the succession of mountain ranges and intermontane basins which characterises the modern morphology of the region. In this thesis, new quantitative constraints on the exhumation histories of several mountain ranges have been obtained by using low temperature thermochronological data from 95 samples (zircon (U-Th)/He, apatite fission track and (U-Th)/He). Time-temperature histories derived by modelling the thermochronologic data of individual samples identify at least two stages of Cenozoic cooling in most of the region’s mountain ranges: (1) initially low cooling rates (<1°C/Myr) during the tectonic quiescent period and (2) increased cooling in the late Cenozoic, which occurred diachronously and with variable magnitude in different ranges. This second cooling stage is interpreted to represent increased erosion caused by active deformation, and in many of the sampled mountain ranges, provides the first available constraints on the timing of late Cenozoic deformation. New constraints on the timing of deformation have also been derived from the sedimentary record of intermontane basins. In the intermontane Issyk Kul basin, new magnetostratigraphic data from two sedimentary sections suggests that deposition of the first Cenozoic syn-tectonic sediments commenced at ~26 Ma. Zircon U-Pb provenance data, paleocurrent and conglomerate clast analysis reveals that these sediments were sourced from the Terskey Range to the south of the basin, suggesting that the onset of the late Cenozoic deformation occurred >26 Ma in that particular range. Elsewhere, growth strata relationships are used to identify syn-tecotnic deposition and constrain the timing of nearby deformation. Collectively, these new constraints obtained from thermochronologic and sedimentary data have allowed me to infer the spatiotemporal distribution of deformation in a transect through the Central Kyrgyz Tien Shan, and determine the order in which mountain ranges started deforming. These data suggest that deformation began in a few widely-spaced mountain ranges in the late Oligocene and early Miocene. Typically, these earlier mountain ranges are bounded on at least one side by a reactivated structure, which probably corresponds to the frictionally weakest and most suitably orientated inherited structures for accommodating the roughly north-south directed horizontal crustal shortening of the late Cenozoic. Moreover, tectonically-induced rock uplift in the Terskey Range, following the reactivation of the bounding structure before 26 Ma, likely caused significant surface uplift across the range, which in turn lead to enhanced orographic precipitation. These wetter conditions have been inferred from stable isotope data collected in the two magnetostratigraphically-dated sections in the Issyk Kul basin. Subsequently, in the late Miocene (~12‒5 Ma), more mountain ranges and inherited structures appear to have started actively deforming. Importantly, the onset of deformation at these locations in the late Miocene coincides with an increase in exhumation of ranges that had started deforming earlier in the late Oligocene‒early Miocene. Based on this observation, I have suggested that there must have been an overall increase in the rate of horizontal crustal shortening across the Central Kyrgyz Tien Shan, which likely relates to regional tectonic changes that affected much of Central Asia. Many of the mountain ranges that started deforming in the late Miocene were associated with out-of-sequence tectonic reactivation and initiation, which lead to the partitioning of larger intermontane basins. Moreover, within most of the intermontane basins in the Central Kyrgyz Tien Shan, this inferred late Miocene increase in horizontal crustal shortening occurs roughly at the same time as an increase in sedimentation rates and a significant change sediment composition. Therefore, I have suggested that the overall magnitude of deformational processes increased in the late Miocene, promoting more flexural subsidence in the intermontane basins of the Central Kyrgyz Tien Shan.
The contribution of the warm-hot intergalactic medium to the CMB anisotropies and distortions
(2013)
Tetrahalidocuprat(II)-komplexe : Untersuchungen zur Relation von Struktur- und EPR-Parametern
(2013)
The correction of software failures tends to be very cost-intensive because their debugging is an often time-consuming development activity. During this activity, developers largely attempt to understand what causes failures: Starting with a test case that reproduces the observable failure they have to follow failure causes on the infection chain back to the root cause (defect). This idealized procedure requires deep knowledge of the system and its behavior because failures and defects can be far apart from each other. Unfortunately, common debugging tools are inadequate for systematically investigating such infection chains in detail. Thus, developers have to rely primarily on their intuition and the localization of failure causes is not time-efficient. To prevent debugging by disorganized trial and error, experienced developers apply the scientific method and its systematic hypothesis-testing. However, even when using the scientific method, the search for failure causes can still be a laborious task. First, lacking expertise about the system makes it hard to understand incorrect behavior and to create reasonable hypotheses. Second, contemporary debugging approaches provide no or only partial support for the scientific method. In this dissertation, we present test-driven fault navigation as a debugging guide for localizing reproducible failures with the scientific method. Based on the analysis of passing and failing test cases, we reveal anomalies and integrate them into a breadth-first search that leads developers to defects. This systematic search consists of four specific navigation techniques that together support the creation, evaluation, and refinement of failure cause hypotheses for the scientific method. First, structure navigation localizes suspicious system parts and restricts the initial search space. Second, team navigation recommends experienced developers for helping with failures. Third, behavior navigation allows developers to follow emphasized infection chains back to root causes. Fourth, state navigation identifies corrupted state and reveals parts of the infection chain automatically. We implement test-driven fault navigation in our Path Tools framework for the Squeak/Smalltalk development environment and limit its computation cost with the help of our incremental dynamic analysis. This lightweight dynamic analysis ensures an immediate debugging experience with our tools by splitting the run-time overhead over multiple test runs depending on developers’ needs. Hence, our test-driven fault navigation in combination with our incremental dynamic analysis answers important questions in a short time: where to start debugging, who understands failure causes best, what happened before failures, and which state properties are infected.
Famously, Einstein read off the geometry of spacetime from Maxwell's equations. Today, we take this geometry that serious that our fundamental theory of matter, the standard model of particle physics, is based on it. However, it seems that there is a gap in our understanding if it comes to the physics outside of the solar system. Independent surveys show that we need concepts like dark matter and dark energy to make our models fit with the observations. But these concepts do not fit in the standard model of particle physics. To overcome this problem, at least, we have to be open to matter fields with kinematics and dynamics beyond the standard model. But these matter fields might then very well correspond to different spacetime geometries. This is the basis of this thesis: it studies the underlying spacetime geometries and ventures into the quantization of those matter fields independently of any background geometry. In the first part of this thesis, conditions are identified that a general tensorial geometry must fulfill to serve as a viable spacetime structure. Kinematics of massless and massive point particles on such geometries are introduced and the physical implications are investigated. Additionally, field equations for massive matter fields are constructed like for example a modified Dirac equation. In the second part, a background independent formulation of quantum field theory, the general boundary formulation, is reviewed. The general boundary formulation is then applied to the Unruh effect as a testing ground and first attempts are made to quantize massive matter fields on tensorial spacetimes.