@phdthesis{Wolf2021, author = {Wolf, Johannes}, title = {Analysis and visualization of transport infrastructure based on large-scale geospatial mobile mapping data}, doi = {10.25932/publishup-53612}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-536129}, school = {Universit{\"a}t Potsdam}, pages = {vi, 121}, year = {2021}, abstract = {3D point clouds are a universal and discrete digital representation of three-dimensional objects and environments. For geospatial applications, 3D point clouds have become a fundamental type of raw data acquired and generated using various methods and techniques. In particular, 3D point clouds serve as raw data for creating digital twins of the built environment. This thesis concentrates on the research and development of concepts, methods, and techniques for preprocessing, semantically enriching, analyzing, and visualizing 3D point clouds for applications around transport infrastructure. It introduces a collection of preprocessing techniques that aim to harmonize raw 3D point cloud data, such as point density reduction and scan profile detection. Metrics such as, e.g., local density, verticality, and planarity are calculated for later use. One of the key contributions tackles the problem of analyzing and deriving semantic information in 3D point clouds. Three different approaches are investigated: a geometric analysis, a machine learning approach operating on synthetically generated 2D images, and a machine learning approach operating on 3D point clouds without intermediate representation. In the first application case, 2D image classification is applied and evaluated for mobile mapping data focusing on road networks to derive road marking vector data. The second application case investigates how 3D point clouds can be merged with ground-penetrating radar data for a combined visualization and to automatically identify atypical areas in the data. For example, the approach detects pavement regions with developing potholes. The third application case explores the combination of a 3D environment based on 3D point clouds with panoramic imagery to improve visual representation and the detection of 3D objects such as traffic signs. The presented methods were implemented and tested based on software frameworks for 3D point clouds and 3D visualization. In particular, modules for metric computation, classification procedures, and visualization techniques were integrated into a modular pipeline-based C++ research framework for geospatial data processing, extended by Python machine learning scripts. All visualization and analysis techniques scale to large real-world datasets such as road networks of entire cities or railroad networks. The thesis shows that some use cases allow taking advantage of established image vision methods to analyze images rendered from mobile mapping data efficiently. The two presented semantic classification methods working directly on 3D point clouds are use case independent and show similar overall accuracy when compared to each other. While the geometry-based method requires less computation time, the machine learning-based method supports arbitrary semantic classes but requires training the network with ground truth data. Both methods can be used in combination to gradually build this ground truth with manual corrections via a respective annotation tool. This thesis contributes results for IT system engineering of applications, systems, and services that require spatial digital twins of transport infrastructure such as road networks and railroad networks based on 3D point clouds as raw data. It demonstrates the feasibility of fully automated data flows that map captured 3D point clouds to semantically classified models. This provides a key component for seamlessly integrated spatial digital twins in IT solutions that require up-to-date, object-based, and semantically enriched information about the built environment.}, language = {en} } @phdthesis{Richter2018, author = {Richter, Rico}, title = {Concepts and techniques for processing and rendering of massive 3D point clouds}, doi = {10.25932/publishup-42330}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-423304}, school = {Universit{\"a}t Potsdam}, pages = {v, 131}, year = {2018}, abstract = {Remote sensing technology, such as airborne, mobile, or terrestrial laser scanning, and photogrammetric techniques, are fundamental approaches for efficient, automatic creation of digital representations of spatial environments. For example, they allow us to generate 3D point clouds of landscapes, cities, infrastructure networks, and sites. As essential and universal category of geodata, 3D point clouds are used and processed by a growing number of applications, services, and systems such as in the domains of urban planning, landscape architecture, environmental monitoring, disaster management, virtual geographic environments as well as for spatial analysis and simulation. While the acquisition processes for 3D point clouds become more and more reliable and widely-used, applications and systems are faced with more and more 3D point cloud data. In addition, 3D point clouds, by their very nature, are raw data, i.e., they do not contain any structural or semantics information. Many processing strategies common to GIS such as deriving polygon-based 3D models generally do not scale for billions of points. GIS typically reduce data density and precision of 3D point clouds to cope with the sheer amount of data, but that results in a significant loss of valuable information at the same time. This thesis proposes concepts and techniques designed to efficiently store and process massive 3D point clouds. To this end, object-class segmentation approaches are presented to attribute semantics to 3D point clouds, used, for example, to identify building, vegetation, and ground structures and, thus, to enable processing, analyzing, and visualizing 3D point clouds in a more effective and efficient way. Similarly, change detection and updating strategies for 3D point clouds are introduced that allow for reducing storage requirements and incrementally updating 3D point cloud databases. In addition, this thesis presents out-of-core, real-time rendering techniques used to interactively explore 3D point clouds and related analysis results. All techniques have been implemented based on specialized spatial data structures, out-of-core algorithms, and GPU-based processing schemas to cope with massive 3D point clouds having billions of points. All proposed techniques have been evaluated and demonstrated their applicability to the field of geospatial applications and systems, in particular for tasks such as classification, processing, and visualization. Case studies for 3D point clouds of entire cities with up to 80 billion points show that the presented approaches open up new ways to manage and apply large-scale, dense, and time-variant 3D point clouds as required by a rapidly growing number of applications and systems.}, language = {en} } @phdthesis{Semmo2016, author = {Semmo, Amir}, title = {Design and implementation of non-photorealistic rendering techniques for 3D geospatial data}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-99525}, school = {Universit{\"a}t Potsdam}, pages = {XVI, 155}, year = {2016}, abstract = {Geospatial data has become a natural part of a growing number of information systems and services in the economy, society, and people's personal lives. In particular, virtual 3D city and landscape models constitute valuable information sources within a wide variety of applications such as urban planning, navigation, tourist information, and disaster management. Today, these models are often visualized in detail to provide realistic imagery. However, a photorealistic rendering does not automatically lead to high image quality, with respect to an effective information transfer, which requires important or prioritized information to be interactively highlighted in a context-dependent manner. Approaches in non-photorealistic renderings particularly consider a user's task and camera perspective when attempting optimal expression, recognition, and communication of important or prioritized information. However, the design and implementation of non-photorealistic rendering techniques for 3D geospatial data pose a number of challenges, especially when inherently complex geometry, appearance, and thematic data must be processed interactively. Hence, a promising technical foundation is established by the programmable and parallel computing architecture of graphics processing units. This thesis proposes non-photorealistic rendering techniques that enable both the computation and selection of the abstraction level of 3D geospatial model contents according to user interaction and dynamically changing thematic information. To achieve this goal, the techniques integrate with hardware-accelerated rendering pipelines using shader technologies of graphics processing units for real-time image synthesis. The techniques employ principles of artistic rendering, cartographic generalization, and 3D semiotics—unlike photorealistic rendering—to synthesize illustrative renditions of geospatial feature type entities such as water surfaces, buildings, and infrastructure networks. In addition, this thesis contributes a generic system that enables to integrate different graphic styles—photorealistic and non-photorealistic—and provide their seamless transition according to user tasks, camera view, and image resolution. Evaluations of the proposed techniques have demonstrated their significance to the field of geospatial information visualization including topics such as spatial perception, cognition, and mapping. In addition, the applications in illustrative and focus+context visualization have reflected their potential impact on optimizing the information transfer regarding factors such as cognitive load, integration of non-realistic information, visualization of uncertainty, and visualization on small displays.}, language = {en} } @article{BuschmannTrappDoellner2016, author = {Buschmann, Stefan and Trapp, Matthias and D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Animated visualization of spatial-temporal trajectory data for air-traffic analysis}, series = {The Visual Computer}, volume = {32}, journal = {The Visual Computer}, publisher = {Springer}, address = {New York}, issn = {0178-2789}, doi = {10.1007/s00371-015-1185-9}, pages = {371 -- 381}, year = {2016}, abstract = {With increasing numbers of flights worldwide and a continuing rise in airport traffic, air-traffic management is faced with a number of challenges. These include monitoring, reporting, planning, and problem analysis of past and current air traffic, e.g., to identify hotspots, minimize delays, or to optimize sector assignments to air-traffic controllers. To cope with these challenges, cyber worlds can be used for interactive visual analysis and analytical reasoning based on aircraft trajectory data. However, with growing data size and complexity, visualization requires high computational efficiency to process that data within real-time constraints. This paper presents a technique for real-time animated visualization of massive trajectory data. It enables (1) interactive spatio-temporal filtering, (2) generic mapping of trajectory attributes to geometric representations and appearance, and (3) real-time rendering within 3D virtual environments such as virtual 3D airport or 3D city models. Different visualization metaphors can be efficiently built upon this technique such as temporal focus+context, density maps, or overview+detail methods. As a general-purpose visualization technique, it can be applied to general 3D and 3+1D trajectory data, e.g., traffic movement data, geo-referenced networks, or spatio-temporal data, and it supports related visual analytics and data mining tasks within cyber worlds.}, language = {en} } @article{ParedesBooAmoretal.2012, author = {Paredes, E. G. and Boo, M. and Amor, M. and Bruguera, J. D. and D{\"o}llner, J{\"u}rgen Roland Friedrich}, title = {Extended hybrid meshing algorithm for multiresolution terrain models}, series = {International journal of geographical information science}, volume = {26}, journal = {International journal of geographical information science}, number = {5}, publisher = {Routledge, Taylor \& Francis Group}, address = {Abingdon}, issn = {1365-8816}, doi = {10.1080/13658816.2011.615317}, pages = {771 -- 793}, year = {2012}, abstract = {Hybrid terrains are a convenient approach for the representation of digital terrain models, integrating heterogeneous data from different sources. In this article, we present a general, efficient scheme for achieving interactive level-of-detail rendering of hybrid terrain models, without the need for a costly preprocessing or resampling of the original data. The presented method works with hybrid digital terrains combining regular grid data and local high-resolution triangulated irregular networks. Since grid and triangulated irregular network data may belong to different datasets, a straightforward combination of both geometries would lead to meshes with holes and overlapping triangles. Our method generates a single multiresolution model integrating the different parts in a coherent way, by performing an adaptive tessellation of the region between their boundaries. Hence, our solution is one of the few existing approaches for integrating different multiresolution algorithms within the same terrain model, achieving a simple interactive rendering of complex hybrid terrains.}, language = {en} } @phdthesis{Jamil2010, author = {Jamil, Abdlhamed}, title = {Fernerkundung und GIS zur Erfassung, Modellierung und Visualisierung orientalischer Stadtstrukturen : das Beispiel Sanaa (Jemen)}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus-50200}, school = {Universit{\"a}t Potsdam}, year = {2010}, abstract = {Gegenstand dieser Arbeit ist die Konzeption, Entwicklung und exemplarische Implementierung eines generischen Verfahrens zur Erfassung, Verarbeitung, Auswertung und kartographischen Visualisierung urbaner Strukturen im altweltlichen Trockeng{\"u}rtel mittels hochaufl{\"o}sender operationeller Fernerkundungsdaten. Das Verfahren wird am Beispiel der jemenitischen Hauptstadt Sanaa einer Vertreterin des Typus der Orientalischen Stadt angewandt und evaluiert. Das zu entwickelnde Verfahren soll auf Standardverfahren und Systemen der raumbezogenen Informationsverarbeitung basieren und in seinen wesentlichen Prozessschritten automatisiert werden k{\"o}nnen. Daten von hochaufl{\"o}senden operationellen Fernerkundungssystemen (wie z.B. QuickBird, Ikonos u. a.) erlauben die Erkennung und Kartierung urbaner Objekte, wie Geb{\"a}ude, Straßen und sogar Autos. Die mit ihnen erstellten Karten und den daraus gewonnenen Informationen k{\"o}nnen zur Erfassung von Urbanisierungsprozessen (Stadt- und Bev{\"o}lkerungswachstum) herangezogen werden. Sie werden auch zur Generierung von 3D-Stadtmodellen genutzt. Diese dienen z.B. der Visualisierung f{\"u}r touristische Anwendungen, f{\"u}r die Stadtplanung, f{\"u}r L{\"a}rmanalysen oder f{\"u}r die Standortplanung von Mobilfunkantennen. Bei dem in dieser Arbeit erzeugten 3D-Visualisierung wurden jedoch keine Geb{\"a}udedetails erfasst. Entscheidend war vielmehr die Wiedergabe der Siedlungsstruktur, die im Vorhandensein und in der Anordnung der Geb{\"a}ude liegt. In dieser Arbeit wurden Daten des Satellitensensors Quickbird von 2005 verwendet. Sie zeigen einen Ausschnitt der Stadt Sanaa in Jemen. Die Fernerkundungsdaten wurden durch andere Daten, u.a. auch Gel{\"a}ndedaten, erg{\"a}nzt und verifiziert. Das ausgearbeitete Verfahren besteht aus der Klassifikation der Satellitenbild-aufnahme, die u.a. pixelbezogen und f{\"u}r jede Klasse einzeln (pixelbezogene Klassifikation auf Klassenebene) durchgef{\"u}hrt wurde. Zus{\"a}tzlich fand eine visuelle Interpretation der Satellitenbildaufnahme statt, bei der einzelne Fl{\"a}chen und die Straßen digitalisiert und die Objekte mit Symbolen gekennzeichnet wurden. Die aus beiden Verfahren erstellten Stadtkarten wurden zu einer fusioniert. Durch die Kombination der Ergebnisse werden die Vorteile beider Karten in einer vereint und ihre jeweiligen Schw{\"a}chen beseitigt bzw. minimiert. Die digitale Erfassung der Konturlinien auf der Orthophotomap von Sanaa erlaubte die Erstellung eines Digitalen Gel{\"a}ndemodells, das der dreidimensionalen Darstellung des Altstadtbereichs von Sanaa diente. Die 3D-Visualisierung wurde sowohl von den pixelbezogenen Klassifikationsergebnissen auf Klassenebene als auch von der digitalen Erfassung der Objekte erstellt. Die Ergebnisse beider Visualisierungen wurden im Anschluss in einer Stadtkarte vereint. Bei allen Klassifikationsverfahren wurden die asphaltierten Straßen, die Vegetation und einzeln stehende Geb{\"a}ude sehr gut erfasst. Die Klassifikation der Altstadt gestaltete sich aufgrund der dort f{\"u}r die Klassifikation herrschenden ung{\"u}nstigen Bedingungen am problematischsten. Die insgesamt besten Ergebnisse mit den h{\"o}chsten Genauigkeitswerten wurden bei der pixelbezogenen Klassifikation auf Klassenebene erzielt. Dadurch, dass jede Klasse einzeln klassifiziert wurde, konnte die zu einer Klasse geh{\"o}rende Fl{\"a}che besser erfasst und nachbearbeitet werden. Die Datenmenge wurde reduziert, die Bearbeitungszeit somit k{\"u}rzer und die Speicherkapazit{\"a}t geringer. Die Auswertung bzw. visuelle Validierung der pixel-bezogenen Klassifikationsergebnisse auf Klassenebene mit dem Originalsatelliten-bild gestaltete sich einfacher und erfolgte genauer als bei den anderen durch-gef{\"u}hrten Klassifikationsverfahren. Außerdem war es durch die alleinige Erfassung der Klasse Geb{\"a}ude m{\"o}glich, eine 3D-Visualisierung zu erzeugen. Bei einem Vergleich der erstellten Stadtkarten ergibt sich, dass die durch die visuelle Interpretation erstellte Karte mehr Informationen enth{\"a}lt. Die von den pixelbezogenen Klassifikationsergebnissen auf Klassenebene erstellte Karte ist aber weniger arbeits- und zeitaufwendig zu erzeugen. Zudem arbeitet sie die Struktur einer orientalischen Stadt mit den wesentlichen Merkmalen besser heraus. Durch die auf Basis der 2D-Stadtkarten erstellte 3D-Visualisierung wird ein anderer r{\"a}umlicher Eindruck vermittelt und bestimmte Elemente einer orientalischen Stadt deutlich gemacht. Dazu z{\"a}hlen die sich in der Altstadt befindenden Sackgassen und die ehemalige Stadtmauer. Auch die f{\"u}r Sanaa typischen Hochh{\"a}user werden in der 3D-Visualisierung erkannt. Insgesamt wurde in der Arbeit ein generisches Verfahren entwickelt, dass mit geringen Modifikationen auch auf andere st{\"a}dtische R{\"a}ume des Typus orientalische Stadt angewendet werden kann.}, language = {de} }