Institut für Informatik und Computational Science
Refine
Has Fulltext
- no (714) (remove)
Year of publication
Document Type
- Article (531)
- Monograph/Edited Volume (82)
- Doctoral Thesis (64)
- Other (26)
- Part of a Book (4)
- Preprint (4)
- Conference Proceeding (3)
Language
- English (714) (remove)
Is part of the Bibliography
- yes (714) (remove)
Keywords
- Answer set programming (10)
- answer set programming (8)
- Answer Set Programming (7)
- Machine learning (3)
- formal languages (3)
- monitoring (3)
- security (3)
- Analytical models (2)
- Automata systems (2)
- E-learning (2)
- Equilibrium logic (2)
- Event mapping (2)
- Fault tolerance (2)
- Internet of Things (2)
- Lindenmayer systems (2)
- MQTT (2)
- Machine Learning (2)
- Modeling (2)
- Non-monotonic reasoning (2)
- Optimization (2)
- Parameterized complexity (2)
- Preference Handling (2)
- Process mining (2)
- ResNet (2)
- Theory (2)
- bioinformatics (2)
- cooperating systems (2)
- online learning (2)
- radhard design (2)
- reliability (2)
- verification (2)
- (FPGA) (1)
- (SET) count rate (1)
- 2-tag system (1)
- 3D modeling (1)
- 3D visualization (1)
- AODV (1)
- ASIC (1)
- Absorbed dose (1)
- Abstraction (1)
- Access control (1)
- Active evaluation (1)
- Ad hoc routing (1)
- Adaptivity (1)
- Advanced Video Codec (AVC) (1)
- Aggregates (1)
- Algorithm configuration (1)
- Algorithm portfolios (1)
- Algorithms (1)
- Android Security (1)
- Android hybrid apps (1)
- Animal building (1)
- Answer Set Solving modulo Theories (1)
- Anti-cancer drugs (1)
- Argumentation structure (1)
- Artificial Neuronal Network (1)
- Augmentation (1)
- Augmented and virtual reality (1)
- Automated parallelization (1)
- Automatically controlled windows (1)
- Backdoors (1)
- Batch processing (1)
- Bean (1)
- Benchmark testing; (1)
- Blind users (1)
- Boolean logic models (1)
- Bot Detection (1)
- Business Process (1)
- Business process intelligence (1)
- CP-Logic (1)
- Campus (1)
- Circuit faults (1)
- Clock tree (1)
- Cloud (1)
- Cluster Computing (1)
- Cluster computing (1)
- Code generation (1)
- Coherent phonons (1)
- Combinatorial multi-objective optimization (1)
- Complex optimization (1)
- Complexity (1)
- Computational complexity (1)
- Computational grid (1)
- Computer security (1)
- Computing with DNA (1)
- Conformant Planning (1)
- Conrad Hal Waddington (1)
- Constraint satisfaction (1)
- Context awareness (1)
- Contextualized learning (1)
- Continuous Testing (1)
- Continuous Versioning (1)
- Convolution (1)
- Course timetabling (1)
- Customer ownership (1)
- D-galactosamine (1)
- DMR (1)
- DNA hairpin formation (1)
- DRMAA (1)
- DRMS (1)
- Data federation (1)
- Database (1)
- Deal of the Day (1)
- Debugging (1)
- Decidability (1)
- Declarative Problem Solving (1)
- Declare (1)
- Deep Learning (1)
- Deep learning (1)
- Denotational semantics (1)
- Design (1)
- Design for testability (DFT) (1)
- Digital image analysis (1)
- Digitalization (1)
- Dose rate (1)
- Double cell upsets (DCUs) (1)
- Dynamical X-ray theory (1)
- E-teaching (1)
- EDC (1)
- EEG (1)
- Edge Computing (1)
- Educational game (1)
- Educational timetabling (1)
- Encoding (1)
- Engines (1)
- Enterprise Architecture (1)
- Entity Linking (1)
- Epigenetic landscape (1)
- Epistemic Logic Programs (1)
- Erklärbarkeit (1)
- Evaluation (1)
- Evolution (1)
- Experimentation (1)
- Explainability (1)
- Explicit negation (1)
- Explore-first Programming (1)
- Extensibility (1)
- Extreme Model-Driven Development (1)
- FEDC (1)
- FPGA (1)
- Fault Localization (1)
- Fault tolerant systems (1)
- Feature extraction (1)
- Field programmable gate arrays (1)
- Finite automata (1)
- Flip-flops (1)
- Forgetting (1)
- Framework (1)
- Freshmen (1)
- GERBIL (1)
- Geschäftsprozess (1)
- Gesture input (1)
- Graph Convolutional Neural Networks (1)
- Graph Embedding (1)
- Green computing (1)
- Grounded theory (1)
- Grounding Theory (1)
- H.264 (1)
- Hairpin completions (1)
- Hairpin reductions (1)
- Hardware accelerator (1)
- Heat diffusion (1)
- Hierarchically configurable mask register (1)
- Histograms (1)
- Https traffic (1)
- Human Factors (1)
- Hurricane Sandy (1)
- Hybrid App (1)
- IaaS (1)
- Identifiers (1)
- Image and video stylization (1)
- Image resolution (1)
- Imperative calculi (1)
- Improving classroom (1)
- Incoherent phonons (1)
- Incremental answer set programming (1)
- Industrie 4.0 (1)
- Industry 4.0 (1)
- Inference (1)
- Information federation (1)
- Information integration (1)
- Information retrieval (1)
- Information security (1)
- Insurance industry (1)
- Integrated circuit modeling (1)
- Internet (1)
- Interpretability (1)
- Interpretierbarkeit (1)
- Job monitoring (1)
- Job submission (1)
- Kernel (1)
- Kernelization (1)
- Key input (1)
- Knowledge (1)
- Knowledge Management (1)
- Knowledge representation (1)
- Künstliche Neuronale Netzwerke (1)
- L systems (1)
- LBA problem (1)
- Landmark visibility (1)
- Learning (1)
- Lernen (1)
- Literature mining (1)
- Liver neoplasms (1)
- Load Balancing (1)
- Localization (1)
- Location awareness (1)
- Logic programming (1)
- Loss (1)
- Low Latency (1)
- Loyalty (1)
- M2M (1)
- Markov processes (1)
- Masking of X-values (1)
- Media in education (1)
- Meta-Programming (1)
- Metric learning (1)
- Minimal perturbation problems (1)
- Mobile Campus Application (1)
- Mobile application (1)
- Mobile devices (1)
- Mobile learning (1)
- Model checking (1)
- Modellierung (1)
- Modelling (1)
- Multi-objective optimization (1)
- Multi-sided platforms (1)
- Multiple interpretation scheme (1)
- N-temperature model (1)
- Nash equilibrium (1)
- Natural language processing (1)
- Natural ventilation (1)
- Network (1)
- Network security (1)
- Neural networks (1)
- Non-Monotonic (1)
- Nonmonotonic reasoning (1)
- OCCI (1)
- OSSE (1)
- Operation problem (1)
- Optimierung (1)
- Parallel SAT solving (1)
- Parallel job execution time estimation (1)
- Particle detector (1)
- Pedagogical issues (1)
- Pedestrian navigation (1)
- Performance Evaluation (1)
- Personal Learning Environment (1)
- Personalization (1)
- Pervasive computing (1)
- Pervasive game (1)
- Pervasive learning (1)
- Phantoms (1)
- Planar tactile display (1)
- Plant identification (1)
- Polarization (1)
- Privacy Protection (1)
- Process (1)
- Process Management (1)
- Process model analysis (1)
- Product lifecycle management (1)
- Programming (1)
- Programming by optimization (1)
- Prototyping (1)
- Prozess (1)
- Prozessmanagement (1)
- RADFET (1)
- RADFETs (1)
- REST (1)
- RSA triangle (1)
- Radiation hardness (1)
- Random access memory (1)
- Ranking (1)
- Reasoning (1)
- Region of Interest (1)
- Reproducibility of results (1)
- Reversibility (1)
- SET pulsewidth distribution (1)
- SOA (1)
- SWOT (1)
- Scalability (1)
- Scale-invariant feature transform (SIFT) (1)
- Scientific images (1)
- Screen reader (1)
- Seamless learning (1)
- Search problems (1)
- Security (1)
- Self-adaptive MPSoC (1)
- Semantic Interoperability (1)
- Semantic data (1)
- Semantic web (1)
- Semilinearity property (1)
- Sequence embeddings (1)
- Service orientation (1)
- Service-oriented Architecture (1)
- Service-oriented architecture (1)
- Sharing (1)
- Signal processing (1)
- Signaling transduction networks (1)
- Simulation (1)
- Simulations (1)
- Single event effect (1)
- Single event upsets (1)
- Single-event transient (SET) (1)
- Social Media Analysis (1)
- Splicing (1)
- Splicing processor (1)
- Static Analysis (1)
- Statistical relational learning (1)
- Stochastic relational process (1)
- Strong equivalence (1)
- Structural equation modeling (1)
- Systems biology (1)
- Systems of parallel communicating (1)
- TMR (1)
- Temporal Answer Set Solving (1)
- Theory formation (1)
- Thermoelasticity (1)
- Time series (1)
- Tomography (1)
- Tools (1)
- Tracking (1)
- Traffic data (1)
- Tree decomposition (1)
- Treewidth (1)
- Treewidth-aware reductions (1)
- Triple modular redundancy (TMR) (1)
- Tumor types (1)
- Turing machine (1)
- Type and effect systems (1)
- UAV imagery (1)
- UX (1)
- Ubiquitous learning (1)
- Ultrafast dynamics (1)
- Unary languages (1)
- Uniform Access Principle (1)
- University Service Bus (1)
- Usability testing (1)
- User submission pattern (1)
- User-centred design (1)
- VGG16 (1)
- Value network (1)
- Verification (1)
- Virtual reality (1)
- Visual metaphor (1)
- Wireless Sensor Networks (1)
- Wissen (1)
- Wissensmanagement (1)
- Word embeddings (1)
- X-masking (1)
- X-ray computed (1)
- X-values (1)
- accepting grammars (1)
- action and change (1)
- activities (1)
- activity (1)
- acute liver failure (1)
- acyclicity properties (1)
- adversarial classification (1)
- algorithm schedules (1)
- algorithms (1)
- analysis (1)
- annealing (1)
- anti-cancer drugs (1)
- anxiety (1)
- approximate model counting (1)
- architecture (1)
- argument mining (1)
- arousal (1)
- artistic rendering (1)
- asynchrounous design (1)
- authentication (1)
- autism (1)
- automata (1)
- automated driving (1)
- automated guided vehicle routing (1)
- automated planning (1)
- automatic feedback (1)
- behavioral (1)
- behavioral abstraction (1)
- belief merging (1)
- belief revision (1)
- benchmark (1)
- bibliometric analysis (1)
- block representation (1)
- bootstrapping (1)
- brain-computer interface (1)
- bundled data (1)
- camera sensor (1)
- car assembly operations (1)
- cellular automata (1)
- circuit Faults (1)
- citation analysis (1)
- click controller (1)
- clocks (1)
- co-citation analysis (1)
- co-occurrence analysis (1)
- code generation (1)
- coherence relation (1)
- collaborative learning (1)
- combinatorial optimization problems (1)
- combined task and motion planning (1)
- common spatial patterns (1)
- competition (1)
- complex networks (1)
- compliance (1)
- computer science education (1)
- computer vision (1)
- concession (1)
- concurrent checking (1)
- conductive argument (1)
- connective (1)
- connectivity (1)
- consistency (1)
- consistency checking (1)
- consistency measures (1)
- context-free grammar (1)
- context-sensitive (1)
- continuous (1)
- contrast (1)
- controlled vocabularies (1)
- corpus analysis (1)
- correlated errors (1)
- course timetabling (1)
- craters (1)
- crop (1)
- decidability questions (1)
- declarative problem solving (1)
- deep learning (1)
- deep neural networks (1)
- deep residual networks (1)
- degree of non-context-freeness (1)
- degree of non-regularity (1)
- degree of non-regulation (1)
- depression (1)
- design flow (1)
- determinism (1)
- detrending (1)
- developmental systems (1)
- diagnosis (1)
- domain-specific APIs (1)
- drug discovery (1)
- drug-sensitivity prediction (1)
- dynamic service binding (1)
- e-learning (1)
- eLectures (1)
- economic ripples (1)
- edge computing (1)
- educational timetabling (1)
- embedded systems (1)
- emission factor (1)
- endothelin (1)
- endothelin-converting enzyme (1)
- ensemble kalman filter (1)
- ensemble methods (1)
- error propagation (1)
- evaluation (1)
- event-related desynchronization (1)
- evolution (1)
- external ambiguity (1)
- extreme weather (1)
- face tracking (1)
- facial expression (1)
- fading (1)
- fault tolerance (1)
- field-programmable gate array (1)
- finite model computation (1)
- finite state sequential transducers (1)
- firmware update (1)
- formal (1)
- formal argumentation systems (1)
- functions (1)
- gait (1)
- gap-filling (1)
- geovisualization (1)
- gradient boosting (1)
- grammar (1)
- graph analysis (1)
- greenhouse gas (1)
- hardware accelerator (1)
- hardware architecture (1)
- higher education (1)
- hybrid solving (1)
- ice harboring (1)
- image classification (1)
- image processing (1)
- image recognition (1)
- imaging (1)
- impacts (1)
- incremental SVM (1)
- informal logic (1)
- information flow control (1)
- internal ambiguity (1)
- intrusion detection (1)
- irradiation (1)
- joint lab (1)
- kidney cancer (1)
- knowledge representation and nonmonotonic reasoning (1)
- knowledge representation and reasoning (1)
- latches (1)
- leftmost derivations (1)
- lesson planning (1)
- lesson preparation (1)
- linear programming (1)
- locomotion (1)
- logic programming (1)
- logic-based modeling (1)
- loop formulas (1)
- loose programming (1)
- loss propagation (1)
- lunar exploration (1)
- machine learning (1)
- machine learning algorithms (1)
- manipulation planning (1)
- measure development (1)
- media (1)
- metabolic network (1)
- metabolism (1)
- metabolomics (1)
- metadata (1)
- metastasis (1)
- natural disasters (1)
- natural language generation (1)
- neighborhood (1)
- neural networks (1)
- neutral endopeptidase (1)
- nonphotorealistic rendering (NPR) (1)
- o-ambiguity (1)
- on-farm evaluation (1)
- oneM2M Ontology (1)
- pMOS radiation dosimeter (1)
- parallel processing (1)
- parallel rewriting (1)
- parity aggregate operator (1)
- parsing (1)
- pdf forms (1)
- perception (1)
- perception differences (1)
- physical computing (1)
- planning (1)
- plug-ins (1)
- policy evaluation (1)
- portfolio-based solving (1)
- predictive models (1)
- premise acceptability (1)
- process model alignment (1)
- process modeling (1)
- program encodings (1)
- programmed grammars (1)
- projection (1)
- proof complexity (1)
- pruritus (1)
- pulse stretching inverters (1)
- quality of life (1)
- quantum (1)
- random forest (1)
- real arguments (1)
- real-time (1)
- real-time mapping (1)
- real-walking (1)
- reference (1)
- referential effectiveness (1)
- regression (1)
- regular language (1)
- relevance (1)
- reliability analysis (1)
- resources (1)
- restricted parallelism (1)
- risk analysis (1)
- safety (1)
- satisfiability (1)
- selective fault tolerance (1)
- self-adaptive multiprocessing system (1)
- self-checking (1)
- semantic web (1)
- sensitivity (1)
- simplicity (1)
- single event upset (1)
- single event upsets (1)
- single-event transient (1)
- single-trial-analysis (1)
- site-specific weed management (1)
- sleep quality (1)
- smart farming (1)
- smartphone (1)
- soft errors (1)
- solar particle event (1)
- space missions (1)
- stable model semantics (1)
- state complexity (1)
- static analysis (1)
- static prediction games (1)
- strong equivalence (1)
- sufficiency (1)
- suicidal ideations (1)
- supply chains (1)
- support system (1)
- support vector machines (1)
- tableau calculi (1)
- teacher training (1)
- teaching (1)
- technical notes and rapid communications (1)
- tele-teaching (1)
- test response compaction (1)
- theory of computation (1)
- timing (1)
- tools (1)
- transient Faults (1)
- transient analysis (1)
- triangulated irregular networks (1)
- triple modular redundancy (1)
- unfounded sets (1)
- user experience (1)
- video annotation (1)
- virtual mobility (1)
- wheat crops (1)
- work productivity (1)
- yellow rust (1)
The introduction of columnar in-memory databases, along with hardware evolution, has made the execution of transactional and analytical enterprise application workloads on a single system both feasible and viable. Yet, we argue that executing analytical aggregate queries directly on the transactional data can decrease the overall system performance. Despite the aggregation capabilities of columnar in-memory databases, the direct access to records of a materialized aggregate is always more efficient than aggregating on the fly. The traditional approach to materialized aggregates, however, introduces significant overhead in terms of materialized view selection, maintenance, and exploitation. When this overhead is handled by the application, it increases the application complexity, and can slow down the transactional throughput of inserts, updates, and deletes.
In this thesis, we motivate, propose, and evaluate the aggregate cache, a materialized aggregate engine in the main-delta architecture of a columnar in-memory database that provides efficient means to handle costly aggregate queries of enterprise applications. For our design, we leverage the specifics of the main-delta architecture that separates a table into a main and delta partition. The central concept is to only cache the partial aggregate query result as defined on the main partition of a table, because the main partition is relatively stable as records are only inserted into the delta partition. We contribute by proposing incremental aggregate maintenance and query compensation techniques for mixed workloads of enterprise applications. In addition, we introduce aggregate profit metrics that increase the likelihood of persisting the most profitable aggregates in the aggregate cache.
Query compensation and maintenance of materialized aggregates based on joins of multiple tables is expensive due to the partitioned tables in the main-delta architecture. Our analysis of enterprise applications has revealed several data schema and workload patterns. This includes the observation that transactional data is persisted in header and item tables, whereas in many cases, the insertion of related header and item records is executed in a single database transaction. We contribute by proposing an approach to transport these application object semantics to the database system and optimize the query processing using the aggregate cache by applying partition pruning and predicate pushdown techniques.
For the experimental evaluation, we propose the FICO benchmark that is based on data from a productive ERP system with extracted mixed workloads. Our evaluation reveals that the aggregate cache can accelerate the execution of aggregate queries up to a factor of 60 whereas the speedup highly depends on the number of aggregated records in the main and delta partitions. In mixed workloads, the proposed aggregate maintenance and query compensation techniques perform up to an order of magnitude better than traditional materialized aggregate maintenance approaches. The introduced aggregate profit metrics outperform existing costbased metrics by up to 20%. Lastly, the join pruning and predicate pushdown techniques can accelerate query execution in the aggregate cache in the presence of multiple partitioned tables by up to an order of magnitude.
In this project I constructed a workflow that takes a DNA sequence as input and provides a phylogenetic tree, consisting of the input sequence and other sequences which were found during a database search. In this phylogenetic tree the sequences are arranged depending on similarities. In bioinformatics, constructing phylogenetic trees is often used to explore the evolutionary relationships of genes or organisms and to understand the mechanisms of evolution itself.
Spotlocator is a game wherein people have to guess the spots of where photos were taken. The photos of a defined area for each game are from panoramio.com. They are published at http://spotlocator. drupalgardens.com with an ID. Everyone can guess the photo spots by sending a special tweet via Twitter that contains the hashtag #spotlocator, the guessed coordinates and the ID of the photo. An evaluation is published for all tweets. The players are informed about the distance to the real photo spots and the positions are shown on a map.
Exploratory Data Analysis
(2014)
In bioinformatics the term exploratory data analysis refers to different methods to get an overview of large biological data sets. Hence, it helps to create a framework for further analysis and hypothesis testing. The workflow facilitates this first important step of the data analysis created by high-throughput technologies. The results are different plots showing the structure of the measurements. The goal of the workflow is the automatization of the exploratory data analysis, but also the flexibility should be guaranteed. The basic tool is the free software R.
The protein classification workflow described in this report enables users to get information about a novel protein sequence automatically. The information is derived by different bioinformatic analysis tools which calculate or predict features of a protein sequence. Also, databases are used to compare the novel sequence with known proteins.
Lessons Learned
(2014)
This chapter summarizes the experience and the lessons we learned concerning the application of the jABC as a framework for design and execution of scientific workflows. It reports experiences from the domain modeling (especially service integration) and workflow design phases and evaluates the resulting models statistically with respect to the SIB library and hierarchy levels.
The Course's SIB Libraries
(2014)
This chapter gives a detailed description of the service framework underlying all the example projects that form the foundation of this book. It describes the different SIB libraries that we made available for the course “Process modeling in the natural sciences” to provide the functionality that was required for the envisaged applications. The students used these SIB libraries to realize their projects.
A major part of the scientific experiments that are carried out today requires thorough computational support. While database and algorithm providers face the problem of bundling resources to create and sustain powerful computation nodes, the users have to deal with combining sets of (remote) services into specific data analysis and transformation processes. Today’s attention to “big data” amplifies the issues of size, heterogeneity, and process-level diversity/integration. In the last decade, especially workflow-based approaches to deal with these processes have enjoyed great popularity. This book concerns a particularly agile and model-driven approach to manage scientific workflows that is based on the XMDD paradigm. In this chapter we explain the scope and purpose of the book, briefly describe the concepts and technologies of the XMDD paradigm, explain the principal differences to related approaches, and outline the structure of the book.
We summarize here the main characteristics and features of the jABC framework, used in the case studies as a graphical tool for modeling scientific processes and workflows. As a comprehensive environment for service-oriented modeling and design according to the XMDD (eXtreme Model-Driven Design) paradigm, the jABC offers much more than the pure modeling capability. Associated technologies and plugins provide in fact means for a rich variety of supporting functionality, such as remote service integration, taxonomical service classification, model execution, model verification, model synthesis, and model compilation. We describe here in short both the essential jABC features and the service integration philosophy followed in the environment. In our work over the last years we have seen that this kind of service definition and provisioning platform has the potential to become a core technology in interdisciplinary service orchestration and technology transfer: Domain experts, like scientists not specially trained in computer science, directly define complex service orchestrations as process models and use efficient and complex domain-specific tools in a simple and intuitive way.
Software-as-a-Service (SaaS) offers several advantages to both service providers and users. Service providers can benefit from the reduction of Total Cost of Ownership (TCO), better scalability, and better resource utilization. On the other hand, users can use the service anywhere and anytime, and minimize upfront investment by following the pay-as-you-go model. Despite the benefits of SaaS, users still have concerns about the security and privacy of their data. Due to the nature of SaaS and the Cloud in general, the data and the computation are beyond the users' control, and hence data security becomes a vital factor in this new paradigm. Furthermore, in multi-tenant SaaS applications, the tenants become more concerned about the confidentiality of their data since several tenants are co-located onto a shared infrastructure.
To address those concerns, we start protecting the data from the provisioning process by controlling how tenants are being placed in the infrastructure. We present a resource allocation algorithm designed to minimize the risk of co-resident tenants called SecPlace. It enables the SaaS provider to control the resource (i.e., database instance) allocation process while taking into account the security of tenants as a requirement.
Due to the design principles of the multi-tenancy model, tenants follow some degree of sharing on both application and infrastructure levels. Thus, strong security-isolation should be present. Therefore, we develop SignedQuery, a technique that prevents one tenant from accessing others' data. We use the Signing Concept to create a signature that is used to sign the tenant's request, then the server can verifies the signature and recognizes the requesting tenant, and hence ensures that the data to be accessed is belonging to the legitimate tenant.
Finally, Data confidentiality remains a critical concern due to the fact that data in the Cloud is out of users' premises, and hence beyond their control. Cryptography is increasingly proposed as a potential approach to address such a challenge. Therefore, we present SecureDB, a system designed to run SQL-based applications over an encrypted database. SecureDB captures the schema design and analyzes it to understand the internal structure of the data (i.e., relationships between the tables and their attributes). Moreover, we determine the appropriate partialhomomorphic encryption scheme for each attribute where computation is possible even when the data is encrypted.
To evaluate our work, we conduct extensive experiments with di↵erent settings. The main use case in our work is a popular open source HRM application, called OrangeHRM. The results show that our multi-layered approach is practical, provides enhanced security and isolation among tenants, and have a moderate complexity in terms of processing encrypted data.
Through the use of next generation sequencing (NGS) technology, a lot of newly sequenced organisms are now available. Annotating those genes is one of the most challenging tasks in sequence biology. Here, we present an automated workflow to find homologue proteins, annotate sequences according to function and create a three-dimensional model.
With the jABC it is possible to realize workflows for numerous questions in different fields. The goal of this project was to create a workflow for the identification of differentially expressed genes. This is of special interest in biology, for it gives the opportunity to get a better insight in cellular changes due to exogenous stress, diseases and so on. With the knowledge that can be derived from the differentially expressed genes in diseased tissues, it becomes possible to find new targets for treatment.