004 Datenverarbeitung; Informatik
Refine
Has Fulltext
- yes (21) (remove)
Year of publication
- 2023 (21) (remove)
Document Type
- Doctoral Thesis (10)
- Monograph/Edited Volume (6)
- Article (3)
- Bachelor Thesis (1)
- Postprint (1)
Language
- English (21) (remove)
Keywords
- Digitalisierung (2)
- Diversity (2)
- digitalization (2)
- machine learning (2)
- 0-day (1)
- 3D Point Clouds (1)
- 3D-Punktwolken (1)
- APT (1)
- Advanced Persistent Threats (1)
- Algorithmic Game Theory (1)
- Algorithmische Spieltheorie (1)
- Anfrageoptimierung (1)
- Anomalieerkennung (1)
- Architekturadaptation (1)
- Augmented reality (1)
- Ausreißererkennung (1)
- Bachelor (1)
- Bahnwesen (1)
- Bedrohungserkennung (1)
- Betriebssysteme (1)
- Bibliometrics (1)
- Bildverarbeitung (1)
- Blockchain (1)
- Blockchains (1)
- Computational Photography (1)
- Course development (1)
- Course marketing (1)
- Courses for female students (1)
- Curricula Development (1)
- Curriculum analysis (1)
- Cyber-Sicherheit (1)
- DBMS (1)
- Data-Mining (1)
- Data-Science (1)
- Datenbank (1)
- Datenbanksysteme (1)
- Datenmodelle (1)
- Dekubitus (1)
- Distributed-Ledger-Technologie (DLT) (1)
- Echtzeit (1)
- Echtzeit-Rendering (1)
- Economics (1)
- Einbruchserkennung (1)
- Endpunktsicherheit (1)
- Ethics (1)
- European Union (1)
- Europäische Union (1)
- Fehlertoleranz (1)
- GPU acceleration (1)
- GPU-Beschleunigung (1)
- Game Dynamics (1)
- Gaussian process state-space models (1)
- Gaussian processes (1)
- Gauß-Prozess Zustandsraummodelle (1)
- Gauß-Prozesse (1)
- Gender (1)
- Generalized Discrimination Networks (1)
- Geschäftsprozessarchitekturen (1)
- Gleichheit (1)
- Graph-Mining (1)
- Graphableitung (1)
- HPI Schul-Cloud (1)
- Hyrise (1)
- Informatics (1)
- Intersectionality (1)
- Karten (1)
- Konnektionskalkül (1)
- Konsensprotokolle (1)
- MERLOT (1)
- Machine-Learning (1)
- Metaverse (1)
- Modelle mit mehreren Versionen (1)
- Netzwerkprotokolle (1)
- Non-photorealistic Rendering (1)
- Omega (1)
- Posenabschätzung (1)
- Preis der Anarchie (1)
- Price of Anarchy (1)
- Quanten-Computing (1)
- Query-Optimierung (1)
- RL (1)
- Real-Time Rendering (1)
- Reverse Engineering (1)
- STEM (1)
- SWIRL (1)
- Schelling Process (1)
- Schelling Prozess (1)
- Schelling Segregation (1)
- Second Life (1)
- Semantic Web (1)
- Sicherheitsanalyse (1)
- Social impact (1)
- Sociotechnical Design (1)
- Software/Hardware Co-Design (1)
- Spieldynamiken (1)
- Sprachlernen im Limes (1)
- Standardisierung (1)
- Systematics (1)
- TPTP (1)
- Taxonomy (1)
- Telemedizin (1)
- Tripel-Graph-Grammatiken (1)
- Verlässlichkeit (1)
- Virtual Reality (1)
- Virtuelle Realität (1)
- Visualisierung (1)
- Visualization (1)
- Web-Based Rendering (1)
- Webbasiertes Rendering (1)
- Women and IT (1)
- advanced persistent threat (1)
- advanced threats (1)
- anomaly detection (1)
- apt (1)
- architectural adaptation (1)
- arithmethische Prozeduren (1)
- arithmetic procedures (1)
- asset management (1)
- automatic theorem prover (1)
- automatisierter Theorembeweiser (1)
- autonomous (1)
- behaviourally correct learning (1)
- bildbasiertes Rendering (1)
- business process architectures (1)
- causal discovery (1)
- causal structure learning (1)
- computational photography (1)
- connection calculus (1)
- consensus protocols (1)
- consistent learning (1)
- convolutional neural networks (1)
- cybersecurity (1)
- data dependencies (1)
- data mining (1)
- data models (1)
- data science (1)
- data-driven (1)
- database (1)
- database optimization (1)
- database systems (1)
- datengetrieben (1)
- decentral identities (1)
- decubitus (1)
- deep Gaussian processes (1)
- deep learning (1)
- dependability (1)
- dezentrale Identitäten (1)
- digital education (1)
- digital sovereignty (1)
- digitale Bildung (1)
- digitale Souveränität (1)
- endpoint security (1)
- equality (1)
- extend (1)
- fault tolerance (1)
- fortschrittliche Angriffe (1)
- functional dependencies (1)
- funktionale Abhängigkeiten (1)
- gefaltete neuronale Netze (1)
- generalized discrimination networks (1)
- geschichtsbewusste Laufzeit-Modelle (1)
- global model management (1)
- globales Modellmanagement (1)
- graph inference (1)
- graph mining (1)
- history-aware runtime models (1)
- image processing (1)
- image stylization (1)
- image-based rendering (1)
- inclusion dependencies (1)
- incremental graph query evaluation (1)
- index selection (1)
- inkrementelle Ausführung von Graphanfragen (1)
- intrusion detection (1)
- juridical recording (1)
- kausale Entdeckung (1)
- kausales Strukturlernen (1)
- konsistentes Lernen (1)
- language learning in the limit (1)
- leanCoP (1)
- maps (1)
- maschinelles Lernen (1)
- model-driven software engineering (1)
- modellgetriebene Softwaretechnik (1)
- multi-version models (1)
- network protocols (1)
- non-photorealistic rendering (1)
- omega (1)
- openHPI (1)
- operating systems (1)
- order dependencies (1)
- outlier detection (1)
- parallel processing (1)
- parallele Verarbeitung (1)
- pose estimation (1)
- probabilistic machine learning (1)
- probabilistisches maschinelles Lernen (1)
- process mining (1)
- quantum computing (1)
- query optimization (1)
- railways (1)
- real-time (1)
- reinforcement learning (1)
- reverse engineering (1)
- security analytics (1)
- selbstbestimmte Identitäten (1)
- self-driving (1)
- self-sovereign identity (1)
- software/hardware co-design (1)
- standardization (1)
- stark verhaltenskorrekt sperrend (1)
- static source-code analysis (1)
- statische Quellcodeanalyse (1)
- strongly behaviourally correct locking (1)
- telemedicine (1)
- temporal graph queries (1)
- temporale Graphanfragen (1)
- threat detection (1)
- tiefe Gauß-Prozesse (1)
- tiefes Lernen (1)
- tptp (1)
- triple graph grammars (1)
- unique column combinations (1)
- unsupervised (1)
- variational inference (1)
- variationelle Inferenz (1)
- verhaltenskorrektes Lernen (1)
- verifiable credentials (1)
- zero-day (1)
- überprüfbare Nachweise (1)
Advancements in computer vision techniques driven by machine learning have facilitated robust and efficient estimation of attributes such as depth, optical flow, albedo, and shading. To encapsulate all such underlying properties associated with images and videos, we evolve the concept of intrinsic images towards intrinsic attributes. Further, rapid hardware growth in the form of high-quality smartphone cameras, readily available depth sensors, mobile GPUs, or dedicated neural processing units have made image and video processing pervasive. In this thesis, we explore the synergies between the above two advancements and propose novel image and video processing techniques and systems based on them. To begin with, we investigate intrinsic image decomposition approaches and analyze how they can be implemented on mobile devices. We propose an approach that considers not only diffuse reflection but also specular reflection; it allows us to decompose an image into specularity, albedo, and shading on a resource constrained system (e.g., smartphones or tablets) using the depth data provided by the built-in depth sensors. In addition, we explore how on-device depth data can further be used to add an immersive dimension to 2D photos, e.g., showcasing parallax effects via 3D photography. In this regard, we develop a novel system for interactive 3D photo generation and stylization on mobile devices. Further, we investigate how adaptive manipulation of baseline-albedo (i.e., chromaticity) can be used for efficient visual enhancement under low-lighting conditions. The proposed technique allows for interactive editing of enhancement settings while achieving improved quality and performance. We analyze the inherent optical flow and temporal noise as intrinsic properties of a video. We further propose two new techniques for applying the above intrinsic attributes for the purpose of consistent video filtering. To this end, we investigate how to remove temporal inconsistencies perceived as flickering artifacts. One of the techniques does not require costly optical flow estimation, while both provide interactive consistency control. Using intrinsic attributes for image and video processing enables new solutions for mobile devices – a pervasive visual computing device – and will facilitate novel applications for Augmented Reality (AR), 3D photography, and video stylization. The proposed low-light enhancement techniques can also improve the accuracy of high-level computer vision tasks (e.g., face detection) under low-light conditions. Finally, our approach for consistent video filtering can extend a wide range of image-based processing for videos.
This technical report presents the results of student projects which were prepared during the lecture “Operating Systems II” offered by the “Operating Systems and Middleware” group at HPI in the Summer term of 2020. The lecture covered ad- vanced aspects of operating system implementation and architecture on topics such as Virtualization, File Systems and Input/Output Systems. In addition to attending the lecture, the participating students were encouraged to gather practical experience by completing a project on a closely related topic over the course of the semester. The results of 10 selected exceptional projects are covered in this report.
The students have completed hands-on projects on the topics of Operating System Design Concepts and Implementation, Hardware/Software Co-Design, Reverse Engineering, Quantum Computing, Static Source-Code Analysis, Operating Systems History, Application Binary Formats and more. It should be recognized that over the course of the semester all of these projects have achieved outstanding results which went far beyond the scope and the expec- tations of the lecture, and we would like to thank all participating students for their commitment and their effort in completing their respective projects, as well as their work on compiling this report.
Ethical issues surrounding modern computing technologies play an increasingly important role in the public debate. Yet, ethics still either doesn’t appear at all or only to a very small extent in computer science degree programs. This paper provides an argument for the value of ethics beyond a pure responsibility perspective and describes the positive value of ethical debate for future computer scientists. It also provides a systematic analysis of the module handbooks of 67 German universities and shows that there is indeed a lack of ethics in computer science education. Finally, we present a principled design of a compulsory course for undergraduate students.
Digital technology offers significant political, economic, and societal opportunities. At the same time, the notion of digital sovereignty has become a leitmotif in German discourse: the state’s capacity to assume its responsibilities and safeguard society’s – and individuals’ – ability to shape the digital transformation in a self-determined way. The education sector is exemplary for the challenge faced by Germany, and indeed Europe, of harnessing the benefits of digital technology while navigating concerns around sovereignty. It encompasses education as a core public good, a rapidly growing field of business, and growing pools of highly sensitive personal data. The report describes pathways to mitigating the tension between digitalization and sovereignty at three different levels – state, economy, and individual – through the lens of concrete technical projects in the education sector: the HPI Schul-Cloud (state sovereignty), the MERLOT data spaces (economic sovereignty), and the openHPI platform (individual sovereignty).
A degree course in IT and business administration solely for women (FIW) has been offered since 2009 at the HTW Berlin – University of Applied Sciences. This contribution discusses student motivations for enrolling in such a women only degree course and gives details of our experience over recent years. In particular, the approach to attracting new female students is described and the composition of the intake is discussed. It is shown that the women-only setting together with other factors can attract a new clientele for computer science.
The Security Operations Center (SOC) represents a specialized unit responsible for managing security within enterprises. To aid in its responsibilities, the SOC relies heavily on a Security Information and Event Management (SIEM) system that functions as a centralized repository for all security-related data, providing a comprehensive view of the organization's security posture. Due to the ability to offer such insights, SIEMS are considered indispensable tools facilitating SOC functions, such as monitoring, threat detection, and incident response.
Despite advancements in big data architectures and analytics, most SIEMs fall short of keeping pace. Architecturally, they function merely as log search engines, lacking the support for distributed large-scale analytics. Analytically, they rely on rule-based correlation, neglecting the adoption of more advanced data science and machine learning techniques.
This thesis first proposes a blueprint for next-generation SIEM systems that emphasize distributed processing and multi-layered storage to enable data mining at a big data scale. Next, with the architectural support, it introduces two data mining approaches for advanced threat detection as part of SOC operations.
First, a novel graph mining technique that formulates threat detection within the SIEM system as a large-scale graph mining and inference problem, built on the principles of guilt-by-association and exempt-by-reputation. The approach entails the construction of a Heterogeneous Information Network (HIN) that models shared characteristics and associations among entities extracted from SIEM-related events/logs. Thereon, a novel graph-based inference algorithm is used to infer a node's maliciousness score based on its associations with other entities in the HIN. Second, an innovative outlier detection technique that imitates a SOC analyst's reasoning process to find anomalies/outliers. The approach emphasizes explainability and simplicity, achieved by combining the output of simple context-aware univariate submodels that calculate an outlier score for each entry.
Both approaches were tested in academic and real-world settings, demonstrating high performance when compared to other algorithms as well as practicality alongside a large enterprise's SIEM system.
This thesis establishes the foundation for next-generation SIEM systems that can enhance today's SOCs and facilitate the transition from human-centric to data-driven security operations.
Diversity is a term that is broadly used and challenging for informatics research, development and education. Diversity concerns may relate to unequal participation, knowledge and methodology, curricula, institutional planning etc. For a lot of these areas, measures, guidelines and best practices on diversity awareness exist. A systemic, sustainable impact of diversity measures on informatics is still largely missing. In this paper I explore what working with diversity and gender concepts in informatics entails, what the main challenges are and provide thoughts for improvement. The paper includes definitions of diversity and intersectionality, reflections on the disciplinary basis of informatics and practical implications of integrating diversity in informatics research and development. In the final part, two concepts from the social sciences and the humanities, the notion of “third space”/hybridity and the notion of “feminist ethics of care”, serve as a lens to foster more sustainable ways of working with diversity in informatics.
In this bachelor’s thesis I implement the automatic theorem prover nanoCoP-Ω. This system is the result of porting arithmetic and equality handling procedures first introduced in the automatic theorem prover with arithmetic leanCoP-Ω into the similar system nanoCoP 2.0. To understand these procedures, I first introduce the mathematical background to both automatic theorem proving and arithmetic expressions. I present the predecessor projects leanCoP, nanoCoP and leanCoP-Ω, out of which nanCoP-Ω was developed. This is followed by an extensive description of the concepts the non-clausal connection calculus needed to be extended by, to allow for proving arithmetic expressions and equalities, as well as of their implementation into nanoCoP-Ω. An extensive comparison between both the runtimes and the number of solved problems of the systems nanoCoP-Ω and leanCoP-Ω was made. I come to the conclusion, that nanoCoP-Ω is considerably faster than leanCoP-Ω for small problems, though less well suited for larger problems. Additionally, I was able to construct a non-theorem that nanoCoP-Ω generates a false proof for. I discuss how this pressing issue could be resolved, as well as some possible optimizations and expansions of the system.
In model-driven engineering, the adaptation of large software systems with dynamic structure is enabled by architectural runtime models. Such a model represents an abstract state of the system as a graph of interacting components. Every relevant change in the system is mirrored in the model and triggers an evaluation of model queries, which search the model for structural patterns that should be adapted. This thesis focuses on a type of runtime models where the expressiveness of the model and model queries is extended to capture past changes and their timing. These history-aware models and temporal queries enable more informed decision-making during adaptation, as they support the formulation of requirements on the evolution of the pattern that should be adapted. However, evaluating temporal queries during adaptation poses significant challenges. First, it implies the capability to specify and evaluate requirements on the structure, as well as the ordering and timing in which structural changes occur. Then, query answers have to reflect that the history-aware model represents the architecture of a system whose execution may be ongoing, and thus answers may depend on future changes. Finally, query evaluation needs to be adequately fast and memory-efficient despite the increasing size of the history---especially for models that are altered by numerous, rapid changes.
The thesis presents a query language and a querying approach for the specification and evaluation of temporal queries. These contributions aim to cope with the challenges of evaluating temporal queries at runtime, a prerequisite for history-aware architectural monitoring and adaptation which has not been systematically treated by prior model-based solutions. The distinguishing features of our contributions are: the specification of queries based on a temporal logic which encodes structural patterns as graphs; the provision of formally precise query answers which account for timing constraints and ongoing executions; the incremental evaluation which avoids the re-computation of query answers after each change; and the option to discard history that is no longer relevant to queries. The query evaluation searches the model for occurrences of a pattern whose evolution satisfies a temporal logic formula. Therefore, besides model-driven engineering, another related research community is runtime verification. The approach differs from prior logic-based runtime verification solutions by supporting the representation and querying of structure via graphs and graph queries, respectively, which is more efficient for queries with complex patterns. We present a prototypical implementation of the approach and measure its speed and memory consumption in monitoring and adaptation scenarios from two application domains, with executions of an increasing size. We assess scalability by a comparison to the state-of-the-art from both related research communities. The implementation yields promising results, which pave the way for sophisticated history-aware self-adaptation solutions and indicate that the approach constitutes a highly effective technique for runtime monitoring on an architectural level.
Most machine learning methods provide only point estimates when being queried to predict on new data. This is problematic when the data is corrupted by noise, e.g. from imperfect measurements, or when the queried data point is very different to the data that the machine learning model has been trained with. Probabilistic modelling in machine learning naturally equips predictions with corresponding uncertainty estimates which allows a practitioner to incorporate information about measurement noise into the modelling process and to know when not to trust the predictions. A well-understood, flexible probabilistic framework is provided by Gaussian processes that are ideal as building blocks of probabilistic models. They lend themself naturally to the problem of regression, i.e., being given a set of inputs and corresponding observations and then predicting likely observations for new unseen inputs, and can also be adapted to many more machine learning tasks. However, exactly inferring the optimal parameters of such a Gaussian process model (in a computationally tractable manner) is only possible for regression tasks in small data regimes. Otherwise, approximate inference methods are needed, the most prominent of which is variational inference.
In this dissertation we study models that are composed of Gaussian processes embedded in other models in order to make those more flexible and/or probabilistic. The first example are deep Gaussian processes which can be thought of as a small network of Gaussian processes and which can be employed for flexible regression. The second model class that we study are Gaussian process state-space models. These can be used for time-series modelling, i.e., the task of being given a stream of data ordered by time and then predicting future observations. For both model classes the state-of-the-art approaches offer a trade-off between expressive models and computational properties (e.g. speed or convergence properties) and mostly employ variational inference. Our goal is to improve inference in both models by first getting a deep understanding of the existing methods and then, based on this, to design better inference methods. We achieve this by either exploring the existing trade-offs or by providing general improvements applicable to multiple methods.
We first provide an extensive background, introducing Gaussian processes and their sparse (approximate and efficient) variants. We continue with a description of the models under consideration in this thesis, deep Gaussian processes and Gaussian process state-space models, including detailed derivations and a theoretical comparison of existing methods.
Then we start analysing deep Gaussian processes more closely: Trading off the properties (good optimisation versus expressivity) of state-of-the-art methods in this field, we propose a new variational inference based approach. We then demonstrate experimentally that our new algorithm leads to better calibrated uncertainty estimates than existing methods.
Next, we turn our attention to Gaussian process state-space models, where we closely analyse the theoretical properties of existing methods.The understanding gained in this process leads us to propose a new inference scheme for general Gaussian process state-space models that incorporates effects on multiple time scales. This method is more efficient than previous approaches for long timeseries and outperforms its comparison partners on data sets in which effects on multiple time scales (fast and slowly varying dynamics) are present.
Finally, we propose a new inference approach for Gaussian process state-space models that trades off the properties of state-of-the-art methods in this field. By combining variational inference with another approximate inference method, the Laplace approximation, we design an efficient algorithm that outperforms its comparison partners since it achieves better calibrated uncertainties.