Refine
Year of publication
Document Type
- Article (570) (remove)
Language
- English (570) (remove)
Keywords
- Answer set programming (10)
- answer set programming (8)
- Answer Set Programming (6)
- Computer Science Education (4)
- Competence Measurement (3)
- Machine learning (3)
- Secondary Education (3)
- Theory (3)
- formal languages (3)
- monitoring (3)
- Analytical models (2)
- Automata systems (2)
- Big Data (2)
- Competence Modelling (2)
- Computational thinking (2)
- E-learning (2)
- Equilibrium logic (2)
- Event mapping (2)
- Fault tolerance (2)
- Informatics Education (2)
- Informatics Modelling (2)
- Informatics System Application (2)
- Informatics System Comprehension (2)
- Lindenmayer systems (2)
- Non-monotonic reasoning (2)
- Parameterized complexity (2)
- Process mining (2)
- ResNet (2)
- bioinformatics (2)
- computational thinking (2)
- computer science education (2)
- cooperating systems (2)
- online learning (2)
- radhard design (2)
- relevance (2)
- security (2)
- tools (2)
- (FPGA) (1)
- (SET) count rate (1)
- 2-tag system (1)
- 21st century skills, (1)
- 3D modeling (1)
- 3D visualization (1)
- ABRACADABRA (1)
- AODV (1)
- ASIC (1)
- Abstraction (1)
- Access control (1)
- Achievement (1)
- Active evaluation (1)
- Activity Theory (1)
- Activity-orientated Learning (1)
- Ad hoc routing (1)
- Adaptivity (1)
- Advanced Video Codec (AVC) (1)
- Aggregates (1)
- Algorithm configuration (1)
- Algorithm portfolios (1)
- Algorithms (1)
- Android hybrid apps (1)
- Animal building (1)
- Anti-cancer drugs (1)
- Arduino (1)
- Argumentation structure (1)
- Assessment (1)
- Augmentation (1)
- Augmented and virtual reality (1)
- Austria (1)
- Automated Theorem Proving (1)
- Automated parallelization (1)
- Automatically controlled windows (1)
- Automatisches Beweisen (1)
- Backdoors (1)
- Batch processing (1)
- Bayesian inference (1)
- Bean (1)
- Benchmark testing; (1)
- Blind users (1)
- Bloom’s Taxonomy (1)
- Boolean logic models (1)
- Business process intelligence (1)
- CP-Logic (1)
- CS concepts (1)
- Campus (1)
- Capability approach (1)
- Challenges (1)
- Circuit faults (1)
- Clause Learning (1)
- Clock tree (1)
- Cloud (1)
- Cluster Computing (1)
- Cluster computing (1)
- Code generation (1)
- Cognitive Skills (1)
- Coherent phonons (1)
- Combinatorial multi-objective optimization (1)
- Competences (1)
- Competencies (1)
- Complex optimization (1)
- Complexity (1)
- Computational Thinking (1)
- Computational complexity (1)
- Computational grid (1)
- Computer Science (1)
- Computer Science in Context (1)
- Computer security (1)
- Computing (1)
- Computing with DNA (1)
- Conformant Planning (1)
- Conrad Hal Waddington (1)
- Constraint satisfaction (1)
- Contest (1)
- Context awareness (1)
- Contextualisation (1)
- Contextualized learning (1)
- Continuous Testing (1)
- Continuous Versioning (1)
- Contradictions (1)
- Convolution (1)
- Course timetabling (1)
- Curriculum (1)
- Curriculum Development (1)
- Customer ownership (1)
- D-galactosamine (1)
- DMR (1)
- DNA hairpin formation (1)
- DPLL (1)
- DRMAA (1)
- DRMS (1)
- Data Analysis (1)
- Data Management (1)
- Data Privacy (1)
- Data augmentation (1)
- Data federation (1)
- Database (1)
- Databases (1)
- Deal of the Day (1)
- Debugging (1)
- Decidability (1)
- Declare (1)
- Deep learning (1)
- Defining characteristics of physical computing (1)
- Denotational semantics (1)
- Design (1)
- Design for testability (DFT) (1)
- Digital Competence (1)
- Digital Education (1)
- Digital Revolution (1)
- Digital image analysis (1)
- Digitalization (1)
- Double cell upsets (DCUs) (1)
- Dynamic assessment (1)
- Dynamical X-ray theory (1)
- E-teaching (1)
- EDC (1)
- EEG (1)
- Early Literacy (1)
- Earthquake modeling (1)
- Educational Standards (1)
- Educational game (1)
- Educational software (1)
- Educational timetabling (1)
- Embedded Systems (1)
- Encoding (1)
- Engines (1)
- Entity Linking (1)
- Epigenetic landscape (1)
- Epistemic Logic Programs (1)
- Euclid’s algorithm (1)
- Evaluation (1)
- Evolution (1)
- Experimentation (1)
- Explicit negation (1)
- Explore-first Programming (1)
- Extensibility (1)
- Extreme Model-Driven Development (1)
- FEDC (1)
- FPGA (1)
- Facebook (1)
- Fault Localization (1)
- Fault tolerant systems (1)
- Feature extraction (1)
- Fibonacci numbers (1)
- Field programmable gate arrays (1)
- Finite automata (1)
- Flip-flops (1)
- Forgetting (1)
- Freshmen (1)
- Function (1)
- Fundamental Ideas (1)
- GERBIL (1)
- Gaussian process (1)
- Gesture input (1)
- Graphensuche (1)
- Green computing (1)
- Grounded theory (1)
- H.264 (1)
- Hairpin completions (1)
- Hairpin reductions (1)
- Hardware accelerator (1)
- Hawkes process (1)
- Heat diffusion (1)
- Hierarchically configurable mask register (1)
- Histograms (1)
- Https traffic (1)
- Human Factors (1)
- Hurricane Sandy (1)
- ICT Competence (1)
- ICT competencies (1)
- ICT skills (1)
- IaaS (1)
- Identifiers (1)
- Image and video stylization (1)
- Image resolution (1)
- Imperative calculi (1)
- Improving classroom (1)
- Incoherent phonons (1)
- Incremental answer set programming (1)
- Inference (1)
- Informatics (1)
- Information federation (1)
- Information integration (1)
- Information retrieval (1)
- Information security (1)
- Inquiry-based Learning (1)
- Insurance industry (1)
- Integrated circuit modeling (1)
- Interface design (1)
- Internet of Things (1)
- Job monitoring (1)
- Job submission (1)
- Kernel (1)
- Kernelization (1)
- Key Competencies (1)
- Key input (1)
- Klausellernen (1)
- Knowledge representation (1)
- L systems (1)
- LBA problem (1)
- Landmark visibility (1)
- Learners (1)
- Learning Fields (1)
- Learning ecology (1)
- Learning interfaces development (1)
- Learning with ICT (1)
- Literature mining (1)
- Liver neoplasms (1)
- Load Balancing (1)
- Localization (1)
- Location awareness (1)
- Logarithm (1)
- Logic programming (1)
- Loss (1)
- Low Latency (1)
- Lower Secondary Level (1)
- Loyalty (1)
- MOOCs (1)
- MQTT (1)
- Machine Learning (1)
- Markov processes (1)
- Masking of X-values (1)
- Massive Open Online Courses (1)
- Measurement (1)
- Media in education (1)
- Meta-Programming (1)
- Metric learning (1)
- Minimal perturbation problems (1)
- Mobile application (1)
- Mobile devices (1)
- Mobile learning (1)
- Model checking (1)
- Modeling (1)
- Modelling (1)
- Multi-objective optimization (1)
- Multi-sided platforms (1)
- Multiple interpretation scheme (1)
- Music Technology (1)
- N-temperature model (1)
- NUI (1)
- Nash equilibrium (1)
- Natural Science Education (1)
- Natural language processing (1)
- Natural ventilation (1)
- Network (1)
- Network security (1)
- Neural networks (1)
- NoSQL (1)
- Non-Monotonic (1)
- Nonmonotonic reasoning (1)
- Norway (1)
- Novice programmers (1)
- OCCI (1)
- OSSE (1)
- Operation problem (1)
- Optimization (1)
- Parallel SAT solving (1)
- Parallel job execution time estimation (1)
- Particle detector (1)
- Pedagogical content knowledge (1)
- Pedagogical issues (1)
- Pedestrian navigation (1)
- Performance Evaluation (1)
- Personalization (1)
- Pervasive computing (1)
- Pervasive game (1)
- Pervasive learning (1)
- Phantoms (1)
- Physical Science (1)
- Planar tactile display (1)
- Plant identification (1)
- Polarization (1)
- Preference Handling (1)
- Preprocessing (1)
- Problem Solving (1)
- Process model analysis (1)
- Product lifecycle management (1)
- Programming (1)
- Programming by optimization (1)
- Prototyping (1)
- REST (1)
- RSA triangle (1)
- Random access memory (1)
- Ranking (1)
- Reasoning (1)
- Recommendations for CS-Curricula in Higher Education (1)
- Region of Interest (1)
- Relevanz (1)
- Reproducibility of results (1)
- Reversibility (1)
- SAT (1)
- SET pulsewidth distribution (1)
- SOA (1)
- SWOT (1)
- Sampling (1)
- Scalability (1)
- Scale-invariant feature transform (SIFT) (1)
- Scientific images (1)
- Screen reader (1)
- Seamless learning (1)
- Search problems (1)
- Security (1)
- Self-exciting point process (1)
- Semantic data (1)
- Semantic web (1)
- Semilinearity property (1)
- Sensors (1)
- Sequence embeddings (1)
- Service orientation (1)
- Sharing (1)
- Signal processing (1)
- Signaling transduction networks (1)
- Simulations (1)
- Single event effect (1)
- Single event upsets (1)
- Single-event transient (SET) (1)
- Small Private Online Courses (1)
- Social (1)
- Spatio-temporal ETAS model (1)
- Splicing (1)
- Splicing processor (1)
- Statistical relational learning (1)
- Stochastic relational process (1)
- Strong equivalence (1)
- Structural equation modeling (1)
- Systems biology (1)
- Systems of parallel communicating (1)
- TMR (1)
- Tasks (1)
- Teacher perceptions (1)
- Teachers (1)
- Teaching information security (1)
- Technology proficiency (1)
- Terminology (1)
- Tests (1)
- Theorembeweisen (1)
- Theory formation (1)
- Thermoelasticity (1)
- Time series (1)
- Tomography (1)
- Tools (1)
- Tracking (1)
- Traffic data (1)
- Tree decomposition (1)
- Treewidth (1)
- Treewidth-aware reductions (1)
- Triple modular redundancy (TMR) (1)
- Tumor types (1)
- Turing machine (1)
- Type and effect systems (1)
- UAV imagery (1)
- UX (1)
- Ubiquitous learning (1)
- Ultrafast dynamics (1)
- Unary languages (1)
- Unifikation (1)
- Uniform Access Principle (1)
- Usability testing (1)
- User submission pattern (1)
- User-centred design (1)
- VGG16 (1)
- Value network (1)
- Verification (1)
- Visual metaphor (1)
- Vocational Education (1)
- Wireless Sensor Networks (1)
- Word embeddings (1)
- X-masking (1)
- X-ray computed (1)
- X-values (1)
- Young People (1)
- abstraction (1)
- accepting grammars (1)
- action and change (1)
- activity (1)
- acute liver failure (1)
- acyclicity properties (1)
- adversarial classification (1)
- algorithm schedules (1)
- algorithms (1)
- analogical thinking (1)
- analysis (1)
- anti-cancer drugs (1)
- anxiety (1)
- approximate model counting (1)
- architecture (1)
- argument mining (1)
- arousal (1)
- artistic rendering (1)
- asynchrounous design (1)
- autism (1)
- automata (1)
- automated guided vehicle routing (1)
- automated planning (1)
- automatic feedback (1)
- behavioral abstraction (1)
- belief merging (1)
- belief revision (1)
- benchmark (1)
- bibliometric analysis (1)
- binary representation (1)
- binary search (1)
- block representation (1)
- bootstrapping (1)
- brain-computer interface (1)
- bundled data (1)
- camera sensor (1)
- car assembly operations (1)
- cellular automata (1)
- circuit Faults (1)
- citation analysis (1)
- classroom language (1)
- click controller (1)
- clocks (1)
- co-citation analysis (1)
- co-occurrence analysis (1)
- code generation (1)
- cognitive modifiability (1)
- coherence relation (1)
- collaborative learning (1)
- combinatorial optimization problems (1)
- combined task and motion planning (1)
- common spatial patterns (1)
- competence (1)
- competencies (1)
- competency (1)
- competition (1)
- complexity (1)
- compliance (1)
- comprehension (1)
- computer science teachers (1)
- computer vision (1)
- concession (1)
- concurrent checking (1)
- conductive argument (1)
- connective (1)
- consistency (1)
- consistency checking (1)
- consistency measures (1)
- context-free grammar (1)
- context-sensitive (1)
- contrast (1)
- controlled vocabularies (1)
- corpus analysis (1)
- correlated errors (1)
- course timetabling (1)
- craters (1)
- crop (1)
- cs4fn (1)
- curriculum theory (1)
- decidability questions (1)
- declarative problem solving (1)
- deep learning (1)
- deep neural networks (1)
- deep residual networks (1)
- degree of non-context-freeness (1)
- degree of non-regularity (1)
- degree of non-regulation (1)
- depression (1)
- design flow (1)
- determinism (1)
- detrending (1)
- developmental systems (1)
- diagnosis (1)
- digitally-enabled pedagogies (1)
- divide and conquer (1)
- domain-specific APIs (1)
- drug discovery (1)
- drug-sensitivity prediction (1)
- dynamic service binding (1)
- e-learning (1)
- e-mentoring (1)
- eLectures (1)
- economic ripples (1)
- edge computing (1)
- education (1)
- education and public policy (1)
- educational programming (1)
- educational systems (1)
- educational timetabling (1)
- edutainment (1)
- embedded systems (1)
- emission factor (1)
- endothelin (1)
- endothelin-converting enzyme (1)
- ensemble kalman filter (1)
- ensemble methods (1)
- environments (1)
- error propagation (1)
- evaluation (1)
- event-related desynchronization (1)
- evolution (1)
- exponentiation (1)
- external ambiguity (1)
- extreme weather (1)
- face tracking (1)
- facial expression (1)
- fault tolerance (1)
- field-programmable gate array (1)
- finite model computation (1)
- finite state sequential transducers (1)
- firmware update (1)
- formal (1)
- formal argumentation systems (1)
- fun (1)
- functions (1)
- gap-filling (1)
- geovisualization (1)
- gradient boosting (1)
- grammar (1)
- graph-search (1)
- greenhouse gas (1)
- hardware accelerator (1)
- hardware architecture (1)
- high school (1)
- higher (1)
- higher education (1)
- hybrid solving (1)
- ice harboring (1)
- image classification (1)
- image processing (1)
- image recognition (1)
- imaging (1)
- impacts (1)
- incremental SVM (1)
- informal and formal learning (1)
- informal logic (1)
- informatics education (1)
- information flow control (1)
- innovation (1)
- interactive course (1)
- interactive workshop (1)
- internal ambiguity (1)
- intrusion detection (1)
- key competences in physical computing (1)
- key competencies (1)
- kidney cancer (1)
- kinaesthetic teaching (1)
- knowledge representation and nonmonotonic reasoning (1)
- knowledge representation and reasoning (1)
- latches (1)
- learning (1)
- leftmost derivations (1)
- lesson planning (1)
- lesson preparation (1)
- linear programming (1)
- logic programming (1)
- logic-based modeling (1)
- loop formulas (1)
- loose programming (1)
- loss propagation (1)
- lunar exploration (1)
- machine learning (1)
- machine learning algorithms (1)
- manipulation planning (1)
- measure development (1)
- media (1)
- mediated learning experience (1)
- metabolic network (1)
- metabolism (1)
- metabolomics (1)
- metadata (1)
- metastasis (1)
- mobile learning (1)
- mobile technologies and apps (1)
- natural disasters (1)
- natural language generation (1)
- neighborhood (1)
- networks (1)
- neural networks (1)
- neutral endopeptidase (1)
- nonphotorealistic rendering (NPR) (1)
- o-ambiguity (1)
- on-farm evaluation (1)
- operating system (1)
- organisational evolution (1)
- paper prototyping (1)
- parallel processing (1)
- parallel rewriting (1)
- parameter (1)
- parity aggregate operator (1)
- parsing (1)
- pdf forms (1)
- pedagogy (1)
- perception (1)
- perception differences (1)
- personal (1)
- personal response systems (1)
- philosophical foundation of informatics pedagogy (1)
- physical computing (1)
- physical computing tools (1)
- planning (1)
- plug-ins (1)
- policy evaluation (1)
- portfolio-based solving (1)
- pre-primary level (1)
- predictive models (1)
- premise acceptability (1)
- preprocessing (1)
- primary education (1)
- primary level (1)
- problem-solving (1)
- process model alignment (1)
- process modeling (1)
- professional development (1)
- program encodings (1)
- programmed grammars (1)
- programming (1)
- programming in context (1)
- projection (1)
- proof complexity (1)
- pruritus (1)
- pulse stretching inverters (1)
- quality of life (1)
- quantum (1)
- random forest (1)
- real arguments (1)
- real-time (1)
- real-time mapping (1)
- reference (1)
- referential effectiveness (1)
- regression (1)
- regular language (1)
- reliability (1)
- reliability analysis (1)
- resources (1)
- restricted parallelism (1)
- satisfiability (1)
- secondary computer science education (1)
- secondary education (1)
- selective fault tolerance (1)
- self-adaptive multiprocessing system (1)
- self-checking (1)
- self-efficacy (1)
- semantic web (1)
- simplicity (1)
- single event upset (1)
- single event upsets (1)
- single-event transient (1)
- single-trial-analysis (1)
- site-specific weed management (1)
- sleep quality (1)
- smart farming (1)
- social media (1)
- soft errors (1)
- solar particle event (1)
- space missions (1)
- stable model semantics (1)
- state complexity (1)
- static analysis (1)
- static prediction games (1)
- strong equivalence (1)
- student activation (1)
- student experience (1)
- student perceptions (1)
- students’ conceptions (1)
- students’ knowledge (1)
- sufficiency (1)
- suicidal ideations (1)
- supply chains (1)
- support system (1)
- support vector machines (1)
- tableau calculi (1)
- teacher competencies (1)
- teacher training (1)
- teaching (1)
- teaching informatics in general education (1)
- technical notes and rapid communications (1)
- tele-teaching (1)
- test response compaction (1)
- theorem (1)
- theory of computation (1)
- timing (1)
- tracing (1)
- transient Faults (1)
- transient analysis (1)
- triangulated irregular networks (1)
- triple modular redundancy (1)
- unfounded sets (1)
- unification (1)
- user experience (1)
- user-centred (1)
- verification (1)
- video annotation (1)
- virtual mobility (1)
- virtual reality (1)
- wheat crops (1)
- work productivity (1)
- yellow rust (1)
- ‘unplugged’ computing (1)
Institute
- Institut für Informatik und Computational Science (570) (remove)
In recent years, many efforts have been made to apply image processing techniques for plant leaf identification. However, categorizing leaf images at the cultivar/variety level, because of the very low inter-class variability, is still a challenging task. In this research, we propose an automatic discriminative method based on convolutional neural networks (CNNs) for classifying 12 different cultivars of common beans that belong to three various species. We show that employing advanced loss functions, such as Additive Angular Margin Loss and Large Margin Cosine Loss, instead of the standard softmax loss function for the classification can yield better discrimination between classes and thereby mitigate the problem of low inter-class variability. The method was evaluated by classifying species (level I), cultivars from the same species (level II), and cultivars from different species (level III), based on images from the leaf foreside and backside. The results indicate that the performance of the classification algorithm on the leaf backside image dataset is superior. The maximum mean classification accuracies of 95.86, 91.37 and 86.87% were obtained at the levels I, II and III, respectively. The proposed method outperforms the previous relevant works and provides a reliable approach for plant cultivars identification.
Objectives To compare image quality of deep learning reconstruction (AiCE) for radiomics feature extraction with filtered back projection (FBP), hybrid iterative reconstruction (AIDR 3D), and model-based iterative reconstruction (FIRST). Methods Effects of image reconstruction on radiomics features were investigated using a phantom that realistically mimicked a 65-year-old patient's abdomen with hepatic metastases. The phantom was scanned at 18 doses from 0.2 to 4 mGy, with 20 repeated scans per dose. Images were reconstructed with FBP, AIDR 3D, FIRST, and AiCE. Ninety-three radiomics features were extracted from 24 regions of interest, which were evenly distributed across three tissue classes: normal liver, metastatic core, and metastatic rim. Features were analyzed in terms of their consistent characterization of tissues within the same image (intraclass correlation coefficient >= 0.75), discriminative power (Kruskal-Wallis test p value < 0.05), and repeatability (overall concordance correlation coefficient >= 0.75). Results The median fraction of consistent features across all doses was 6%, 8%, 6%, and 22% with FBP, AIDR 3D, FIRST, and AiCE, respectively. Adequate discriminative power was achieved by 48%, 82%, 84%, and 92% of features, and 52%, 20%, 17%, and 39% of features were repeatable, respectively. Only 5% of features combined consistency, discriminative power, and repeatability with FBP, AIDR 3D, and FIRST versus 13% with AiCE at doses above 1 mGy and 17% at doses >= 3 mGy. AiCE was the only reconstruction technique that enabled extraction of higher-order features. Conclusions AiCE more than doubled the yield of radiomics features at doses typically used clinically. Inconsistent tissue characterization within CT images contributes significantly to the poor stability of radiomics features.
The spatio-temporal epidemic type aftershock sequence (ETAS) model is widely used to describe the self-exciting nature of earthquake occurrences. While traditional inference methods provide only point estimates of the model parameters, we aim at a fully Bayesian treatment of model inference, allowing naturally to incorporate prior knowledge and uncertainty quantification of the resulting estimates. Therefore, we introduce a highly flexible, non-parametric representation for the spatially varying ETAS background intensity through a Gaussian process (GP) prior. Combined with classical triggering functions this results in a new model formulation, namely the GP-ETAS model. We enable tractable and efficient Gibbs sampling by deriving an augmented form of the GP-ETAS inference problem. This novel sampling approach allows us to assess the posterior model variables conditioned on observed earthquake catalogues, i.e., the spatial background intensity and the parameters of the triggering function. Empirical results on two synthetic data sets indicate that GP-ETAS outperforms standard models and thus demonstrate the predictive power for observed earthquake catalogues including uncertainty quantification for the estimated parameters. Finally, a case study for the l'Aquila region, Italy, with the devastating event on 6 April 2009, is presented.
Programs are often subjected to significant optimizing and parallelizing transformations based on extensive dependence analysis. Formal validation of such transformations needs modelling paradigms which can capture both control and data dependences in the program vividly. Being value-based with an inherent scope of capturing parallelism, the untimed coloured Petri net (CPN) models, reported in the literature, fit the bill well; accordingly, they are likely to be more convenient as the intermediate representations (IRs) of both the source and the transformed codes for translation validation than strictly sequential variable-based IRs like sequential control flow graphs (CFGs). In this work, an efficient path-based equivalence checking method for CPN models of programs on integers is presented. Extensive experimentation has been carried out on several sequential and parallel examples. Complexity and correctness issues have been treated rigorously for the method.
This study investigates the use of pulse stretching (skew-sized) inverters for monitoring the variation of count rate and linear energy transfer (LET) of energetic particles. The basic particle detector is a cascade of two pulse stretching inverters, and the required sensing area is obtained by connecting up to 12 two-inverter cells in parallel and employing the required number of parallel arrays. The incident particles are detected as single-event transients (SETs), whereby the SET count rate denotes the particle count rate, while the SET pulsewidth distribution depicts the LET variations. The advantage of the proposed solution is the possibility to sense the LET variations using fully digital processing logic. SPICE simulations conducted on IHP's 130-nm CMOS technology have shown that the SET pulsewidth varies by approximately 550 ps over the LET range from 1 to 100 MeV center dot cm(2) center dot mg(-1). The proposed detector is intended for triggering the fault-tolerant mechanisms within a self-adaptive multiprocessing system employed in space. It can be implemented as a standalone detector or integrated in the same chip with the target system.
The intensity of cosmic radiation may differ over five orders of magnitude within a few hours or days during the Solar Particle Events (SPEs), thus increasing for several orders of magnitude the probability of Single Event Upsets (SEUs) in space-borne electronic systems. Therefore, it is vital to enable the early detection of the SEU rate changes in order to ensure timely activation of dynamic radiation hardening measures. In this paper, an embedded approach for the prediction of SPEs and SRAM SEU rate is presented. The proposed solution combines the real-time SRAM-based SEU monitor, the offline-trained machine learning model and online learning algorithm for the prediction. With respect to the state-of-the-art, our solution brings the following benefits: (1) Use of existing on-chip data storage SRAM as a particle detector, thus minimizing the hardware and power overhead, (2) Prediction of SRAM SEU rate one hour in advance, with the fine-grained hourly tracking of SEU variations during SPEs as well as under normal conditions, (3) Online optimization of the prediction model for enhancing the prediction accuracy during run-time, (4) Negligible cost of hardware accelerator design for the implementation of selected machine learning model and online learning algorithm. The proposed design is intended for a highly dependable and self-adaptive multiprocessing system employed in space applications, allowing to trigger the radiation mitigation mechanisms before the onset of high radiation levels.
In control theory, to solve a finite-horizon sequential decision problem (SDP) commonly means to find a list of decision rules that result in an optimal expected total reward (or cost) when taking a given number of decision steps. SDPs are routinely solved using Bellman's backward induction. Textbook authors (e.g. Bertsekas or Puterman) typically give more or less formal proofs to show that the backward induction algorithm is correct as solution method for deterministic and stochastic SDPs. Botta, Jansson and Ionescu propose a generic framework for finite horizon, monadic SDPs together with a monadic version of backward induction for solving such SDPs. In monadic SDPs, the monad captures a generic notion of uncertainty, while a generic measure function aggregates rewards. In the present paper, we define a notion of correctness for monadic SDPs and identify three conditions that allow us to prove a correctness result for monadic backward induction that is comparable to textbook correctness proofs for ordinary backward induction. The conditions that we impose are fairly general and can be cast in category-theoretical terms using the notion of Eilenberg-Moore algebra. They hold in familiar settings like those of deterministic or stochastic SDPs, but we also give examples in which they fail. Our results show that backward induction can safely be employed for a broader class of SDPs than usually treated in textbooks. However, they also rule out certain instances that were considered admissible in the context of Botta et al. 's generic framework. Our development is formalised in Idris as an extension of the Botta et al. framework and the sources are available as supplementary material.
In this paper, we present a study comparing the depth to diameter (d/D) ratio of small simple craters (200-1000 m) of an area between -88.5 degrees to -90 degrees latitude at the lunar south pole containing Permanent Shadowed Regions (PSRs) versus craters without PSRs. As PSRs can reach temperatures of 110 K and are capable of harboring volatiles, especially water ice, we analyzed the relationship of depth versus diameter ratios and its possible implications for harboring water ice. Variations in the d/D ratios can also be caused by other processes such as degradation, isostatic adjustment, or differences in surface properties. The conducted d/D ratio analysis suggests that a differentiation between craters containing PSRs versus craters without PSRs occurs. Thus, a possible direct relation between d/D ratio, PSRs, and water ice harboring might exist. Our results suggest that differences in the target's surface properties may explain the obtained results. The resulting d/D ratios of craters with PSRs can help to select target areas for future In-Situ Resource Utilization (ISRU) missions.
Machine learning for improvement of thermal conditions inside a hybrid ventilated animal building
(2021)
In buildings with hybrid ventilation, natural ventilation opening positions (windows), mechanical ventilation rates, heating, and cooling are manipulated to maintain desired thermal conditions. The indoor temperature is regulated solely by ventilation (natural and mechanical) when the external conditions are favorable to save external heating and cooling energy. The ventilation parameters are determined by a rule-based control scheme, which is not optimal. This study proposes a methodology to enable real-time optimum control of ventilation parameters. We developed offline prediction models to estimate future thermal conditions from the data collected from building in operation. The developed offline model is then used to find the optimal controllable ventilation parameters in real-time to minimize the setpoint deviation in the building. With the proposed methodology, the experimental building's setpoint deviation improved for 87% of time, on average, by 0.53 degrees C compared to the current deviations.
Deep metric learning employs deep neural networks to embed instances into a metric space such that distances between instances of the same class are small and distances between instances from different classes are large. In most existing deep metric learning techniques, the embedding of an instance is given by a feature vector produced by a deep neural network and Euclidean distance or cosine similarity defines distances between these vectors. This paper studies deep distributional embeddings of sequences, where the embedding of a sequence is given by the distribution of learned deep features across the sequence. The motivation for this is to better capture statistical information about the distribution of patterns within the sequence in the embedding. When embeddings are distributions rather than vectors, measuring distances between embeddings involves comparing their respective distributions. The paper therefore proposes a distance metric based on Wasserstein distances between the distributions and a corresponding loss function for metric learning, which leads to a novel end-to-end trainable embedding model. We empirically observe that distributional embeddings outperform standard vector embeddings and that training with the proposed Wasserstein metric outperforms training with other distance functions.