004 Datenverarbeitung; Informatik
Refine
Has Fulltext
- yes (285)
Year of publication
Document Type
- Journal article (127)
- Doctoral Thesis (80)
- Working Paper (37)
- Preprint (19)
- Conference Proceeding (9)
- Jahresbericht (5)
- Master Thesis (4)
- Report (3)
- Other (1)
Language
- English (257)
- German (27)
- Multiple languages (1)
Keywords
- virtual reality (16)
- Datennetz (14)
- Leistungsbewertung (13)
- Quran (8)
- Robotik (8)
- Koran (7)
- Mobiler Roboter (7)
- Text Mining (7)
- Autonomer Roboter (6)
- Simulation (6)
- Computer Center University of Wuerzburg (5)
- Jahresbericht (5)
- Komplexitätstheorie (5)
- Maschinelles Lernen (5)
- Netzwerk (5)
- Optimierung (5)
- P4 (5)
- Theoretische Informatik (5)
- Visualisierung (5)
- annual report (5)
- artificial intelligence (5)
- machine learning (5)
- Drahtloses Sensorsystem (4)
- Modellierung (4)
- Optimization (4)
- Overlay-Netz (4)
- RZUW (4)
- Routing (4)
- SDN (4)
- Verteiltes System (4)
- XML (4)
- augmented reality (4)
- database (4)
- deep learning (4)
- human-computer interaction (4)
- Algorithmus (3)
- Approximationsalgorithmus (3)
- Bayesian classifier (3)
- Computer Vision (3)
- Computersimulation (3)
- Data Mining (3)
- Deep learning (3)
- Dienstgüte (3)
- Graph (3)
- Graphenzeichnen (3)
- IoT (3)
- Julius-Maximilians-Universität Würzburg (3)
- Komplexität (3)
- Künstliche Intelligenz (3)
- Lokalisation (3)
- Machine Learning (3)
- Mensch-Maschine-Schnittstelle (3)
- Netzwerkmanagement (3)
- Peer-to-Peer-Netz (3)
- Performance Evaluation (3)
- QoE (3)
- Quadrocopter (3)
- Quality of Experience (3)
- Rechenzentrum (3)
- Rechnernetz (3)
- Ressourcenmanagement (3)
- Robotics (3)
- Software Engineering (3)
- Softwarearchitektur (3)
- Textvergleich (3)
- Visualization (3)
- Wissensmanagement (3)
- approximation algorithm (3)
- crowdsensing (3)
- graph drawing (3)
- immersion (3)
- mHealth (3)
- neural networks (3)
- resistance (3)
- simulation (3)
- 5G (2)
- Algorithmische Geometrie (2)
- Ausfallsicheres System (2)
- Ausfallsicherheit (2)
- Base text (2)
- Benutzerschnittstelle (2)
- CSS (2)
- Cascading Style Sheets (2)
- Content Management (2)
- Crowdsourcing (2)
- Dot-Depth Problem (2)
- Drahtloses lokales Netz (2)
- Effizienter Algorithmus (2)
- Entscheidbarkeit (2)
- Entscheidungsfindung (2)
- Fernwartung (2)
- Future Internet (2)
- Gothenburg model (2)
- Human-Robot-Interaction (2)
- IEEE 802.11 (2)
- IT Security (2)
- IT-Sicherheit (2)
- Internet of Things (2)
- Kleinsatellit (2)
- Knowledge Management (2)
- Kreuzung (2)
- Localization (2)
- MP-DCCP (2)
- Maschinelles Sehen (2)
- Mehragentensystem (2)
- Mensch-Maschine-System (2)
- Mensch-Roboter-Interaktion (2)
- Meta-model (2)
- Mixed Reality (2)
- Multimedia (2)
- Mustererkennung (2)
- NP-hardness (2)
- PROLOG <Programmiersprache> (2)
- Programmierbare logische Anordnung (2)
- Punktwolke (2)
- Quadrotor (2)
- Resilience (2)
- Resource Management (2)
- Satellit (2)
- Situation Awareness (2)
- Software Defined Networking (2)
- Teleoperation (2)
- Text mining (2)
- Textual alterations weighting system (2)
- Textual document collation (2)
- Theoretical Computer Science (2)
- User Interface (2)
- Verbotsmuster (2)
- Wissensrepräsentation (2)
- Wrapper <Programmierung> (2)
- XR (2)
- automation (2)
- connected mobility applications (2)
- cosmology (2)
- crossing minimization (2)
- decidability (2)
- design (2)
- dot-depth problem (2)
- education (2)
- educational tool (2)
- endliche Automaten (2)
- endoscopy (2)
- engineering (2)
- evolution (2)
- exposure (2)
- finite automata (2)
- fog computing (2)
- forbidden patterns (2)
- framework (2)
- fully convolutional neural networks (2)
- games (2)
- gastroenterology (2)
- genetics (2)
- historical document analysis (2)
- immersive technologies (2)
- knowledge representation (2)
- measurements (2)
- metabolic modeling (2)
- mobile networks (2)
- mobile robots (2)
- multipath (2)
- multipath scheduling (2)
- natural variation (2)
- navigation (2)
- network calculus (2)
- networks (2)
- ontology (2)
- perception (2)
- prediction (2)
- regular languages (2)
- reguläre Sprachen (2)
- satellite communication (2)
- scalability (2)
- scheduling (2)
- segmentation (2)
- self-aware computing (2)
- sensor (2)
- smart speaker (2)
- spatial presence (2)
- tinnitus (2)
- virtual agent (2)
- virtual environments (2)
- 26S RDNA Data (1)
- 3D Laser Scanning (1)
- 3D Pointcloud (1)
- 3D Punktwolke (1)
- 3D Reconstruction (1)
- 3D Sensor (1)
- 3D Vision (1)
- 3D collation (1)
- 3D fluoroscopy (1)
- 3D point cloud (1)
- 3D thermal mapping (1)
- 3D viewer (1)
- 3D-Rekonstruktion (1)
- 3D-reconstruction methods (1)
- 3DTK toolkit (1)
- 4D-GIS (1)
- 4G Networks (1)
- 5G core network (1)
- 5G-ATSSS (1)
- 5GC (1)
- 6DOF Pose Estimation (1)
- 6G (1)
- ACKR4 (1)
- AI (1)
- AKT (1)
- ATSSSS (1)
- AVA (1)
- Abhängigskeitsgraph (1)
- Ablaufplanung (1)
- Accessibility (1)
- Add-on-Miss (1)
- Admission Control (1)
- Agent <Informatik> (1)
- Agent <Künstliche Intelligenz> (1)
- Agent-based Simulation (1)
- Agentbased System (1)
- Agenten-basierte Simulation (1)
- Analysis (1)
- Anforderungsmanagement (1)
- Angewandte Informatik (1)
- Annotation (1)
- Anwendung (1)
- Anwendungsfall (1)
- Approximation (1)
- Arctic (1)
- Arterie (1)
- Artery (1)
- Aufsatzsammlung (1)
- Aufwandsanalyse (1)
- Automat <Automatentheorie> (1)
- Automata Theory (1)
- Automatentheorie (1)
- Automatisierte Prüfungskorrektur (1)
- Autonomer Agent (1)
- Autonomie (1)
- Autonomous Robot (1)
- Autonomous UAV (1)
- Autonomous multi-vehicle systems (1)
- Autorotation (1)
- BPM (1)
- BPMN (1)
- Backbone-Netz (1)
- Background Knowledge (1)
- Balloon (1)
- Banks Islands (1)
- Barcodes (1)
- Bayes-Klassifikator (1)
- Benutzererlebnis (1)
- Benutzerforschung (1)
- Benutzerinteraktion (1)
- Berechenbarkeit (1)
- Berechnungskomplexität (1)
- Bernoulli stochastics (1)
- Bernoulli-Raum (1)
- Bernoullische Stochastik (1)
- Bernoullispace (1)
- Betriebssystem (1)
- Bewegungsablauf (1)
- Bewegungskompensation (1)
- Bewegungskoordination (1)
- Bewegungsplanung (1)
- Bildverarbeitung (1)
- Biological Networks (1)
- Biology (1)
- Bit Parallelität (1)
- BitTorrent (1)
- Bodenstation (1)
- Boolean Grammar (1)
- Boolean equivalence (1)
- Boolean function (1)
- Boolean functions (1)
- Boolean hierarchy (1)
- Boolean isomorphism (1)
- Boolean tree (1)
- Boolesche Funktionen (1)
- Boolesche Grammatik (1)
- Boolesche Hierarchie (1)
- Brüder Grimm Privatbibliothek (1)
- Business Intelligence (1)
- CASE (1)
- CD4+T cells (1)
- CD8+T cells (1)
- CD95 (1)
- CETCH cycle (1)
- CHI Conference (1)
- CLIP (1)
- CO2-sequestration (1)
- COVID-19 (1)
- Caenorhabditis elegans (1)
- Calibration (1)
- Call Graph (1)
- Causes of revelation (1)
- Chapters arrangement (1)
- Charged aerosol detector (CAD) (1)
- Chord (1)
- Chronology of revelation (1)
- Clones (1)
- Cloud Gaming (1)
- Clustering (1)
- Colonial volvocales chlorophyta (1)
- Communication Networks (1)
- Complex Systems (1)
- Complexity Theory (1)
- Compression (1)
- Computational Geometry (1)
- Computational complexity (1)
- Computer software (1)
- Computerspiel (1)
- Computerunterstütztes Lernen (1)
- Computervirus (1)
- Containerization (1)
- Content Distribution (1)
- Convolutional Neural Network (1)
- Cost Analysis (1)
- Crowdsensing (1)
- CubeSat (1)
- DHT (1)
- DNA (1)
- DNA storage (1)
- Daedalus-Projekt (1)
- Dasycladales chlorophyta (1)
- Databases (1)
- Datenbanken (1)
- Datenbasis (1)
- Datenkommunikationsnetz (1)
- Datenübertragung ; Datensicherung ; Informationstechnik ; Internet ; Computersicherheit (1)
- Deep Georeferencing (1)
- Deep Learning (1)
- Dependency Graph (1)
- Design (1)
- Design and Development (1)
- Dezentrale Regelung (1)
- Diagnosesystem (1)
- Dichotomy (1)
- Dienstleistungen (1)
- Digital Elevation Model (1)
- Digitalisierung (1)
- Dijkstra’s algorithm (1)
- Diskrete Simulation (1)
- Distributed Space Systems (1)
- Domänenspezifische Sprache (1)
- Dot-Depth-Hierarchie (1)
- Drahtloses vermaschtes Netz (1)
- Dreidimensionale Rekonstruktion (1)
- Dreieck (1)
- Dynamic Environments (1)
- Dynamic Memory Management (1)
- Dynamische Speicherverwaltung (1)
- E8 symmetry (1)
- EEG (1)
- EEG frequency band analysis (1)
- EEG preprocessing (1)
- EEG processing (1)
- EPM (1)
- Echtzeitsystem (1)
- Echzeit (1)
- Edge-MEC-Cloud (1)
- Edge-based Intelligence (1)
- Educational Measurement (I2.399) (1)
- Eingebettetes System (1)
- Elasticity tensor (1)
- Elastizitätstensor (1)
- Embedded Systems (1)
- Emotion inference (1)
- Emotionserkennung (1)
- Emotionsinterpretation (1)
- Endnutzer (1)
- Endpoint Mobility (1)
- Energieeffizienz (1)
- Energy efficiency (1)
- Entscheidungsträger (1)
- Erfüllbarkeitsproblem (1)
- Erkennung handschriftlicher Artefakte (1)
- Erweiterte Realität (1)
- Ethik (1)
- Euclidean plane (1)
- Euklidische Ebene (1)
- Expert System (1)
- Expertensystem (1)
- FIFO caching strategies (1)
- FLIMbee (1)
- Fachgespräch (1)
- Fahrsimulation (1)
- Fahrsimulator (1)
- Fairness (1)
- Fallstudie (1)
- Fatty acids (1)
- Feature Based Registration (1)
- Feature-Matching (1)
- Fehlertoleranz (1)
- Feldprogrammierbare Architekturen (1)
- Fernsteuerung (1)
- Field programmable gate array (1)
- Field-programmable Gate Arrays (1)
- Firewall (1)
- Flugkörper (1)
- Forces (1)
- Formale Sprache (1)
- Formation (1)
- Formation Flight (1)
- Formationsbewegung (1)
- Formmessung (1)
- Forschung (1)
- Fragmentation (1)
- Fragmentierung (1)
- Frames (1)
- Frühdruck (1)
- Funkressourcenverwaltung (1)
- GNSS/INS integrated navigation (1)
- Gastroenterologische Endoskopie (1)
- Gay-Array-Bauelement (1)
- Gefühl (1)
- Generalisierung <Kartografie> (1)
- Generation Problem (1)
- Generierungsproblem (1)
- Genetic Optimization (1)
- Genetische Optimierung (1)
- Georeferenzierung (1)
- Gllobal self-localisation (1)
- Globale Selbstlokalisation (1)
- Gothenburg Modell (1)
- Gothenburg model of collation process (1)
- Gradient boosted trees (GBT) (1)
- Graphentheorie (1)
- Grimm brothers personal library (1)
- Ground Station Networks (1)
- H.264 SVC (1)
- H.264/SVC (1)
- HGPS (1)
- HHblits (1)
- HMD (Head-Mounted Display) (1)
- HSPA (1)
- HTML (1)
- HTTP adaptive video streaming (1)
- Halbordnungen (1)
- Handschrift (1)
- Hardware (1)
- Herzkatheter (1)
- Herzkathetereingriff (1)
- Hierarchische Simulation (1)
- High-performance liquid chromatography (HPLC) (1)
- Hintergrundwissen (1)
- Historical Maps (1)
- Historische Karte (1)
- Historische Landkarten (1)
- Hittitology (1)
- Hochschulnetz (1)
- Hospital (1)
- Human-centered computing / Access (1)
- Human-centered computing / Human computer interaction (HCI) / Interaction paradigms / Mixed / augmented reality (1)
- Human-centered computing / Human computer interaction (HCI) / Interaction paradigms / Virtual reality (1)
- Human-centered computing / Human computer interaction (HCI) / Interactiondevices (1)
- Human-centered computing / Human computerinteraction (HCI) / Interaction techniques (1)
- Hurwitz theorem (1)
- I-tasser (1)
- ICEP (1)
- IEEE 802.11e (1)
- IEEE 802.15.4 (1)
- IGFBP2 (1)
- III secretion (1)
- INS/LIDAR integrated navigation (1)
- IP (1)
- IT security (1)
- Image Aesthetic Assessment (1)
- Image Registration (1)
- ImageJ (1)
- Implementierung <Informatik> (1)
- Industrial internet (1)
- Industrie 4.0 (1)
- Inferenz <Künstliche Intelligenz> (1)
- Informatik (1)
- Information Extraction (1)
- Information Retrieval (1)
- Information Visualization (1)
- Information-Retrieval-System (1)
- Innovation Management (1)
- Innovationsmanagement (1)
- Instrument Control Toolbox (1)
- Intelligent Virtual Agents (1)
- InteractionSuitcase (1)
- Interaktion (1)
- Internet (1)
- Internet Protokoll (1)
- Invertierte Liste (1)
- IoT-driven processes (1)
- IronChip Evaluation Package (1)
- Isomorphie (1)
- Itinerare (1)
- Itineraries (1)
- JCAS (1)
- JSF (1)
- Jacobian matrix (1)
- Java 3D (1)
- Java <Programmiersprache> (1)
- Java Frameworks (1)
- Java Message Service (1)
- Julius-Maximilians-Universität Würzburg. Rechenzentrum (1)
- Kademlia (1)
- Kanalzugriff (1)
- Karte (1)
- Kartierung (1)
- Kathará (1)
- Kerneldensity estimation (1)
- Klassendiagramm (1)
- Klassifikation (1)
- Klima (1)
- Knowledge Discovery (1)
- Knowledge Management System (1)
- Knowledge Modeling (1)
- Knowledge representation (1)
- Knowledge-based System (1)
- Knowledge-based Systems Engineering (1)
- Kombinatorik (1)
- Kommunikation (1)
- Kommunikationsnetze (1)
- Komplexes System (1)
- Komplexitätsklasse (1)
- Komplexitätsklasse NP (1)
- Konvexe Zeichnungen (1)
- Konzeptsuche (1)
- Kooperierende mobile Roboter (1)
- Krankenhaus (1)
- Kreuzungsminimierung (1)
- Kryoelektronenmikroskopie (1)
- Kurve (1)
- LC-MS/MS (1)
- LFU (1)
- LRU (1)
- Land Cover Classification (1)
- Land plants (1)
- Landkartenbeschriftung (1)
- Landnutzungskartierung (1)
- Laser scanning (1)
- Lava (1)
- Lawhul-Mahfuz (1)
- Learning (1)
- Lee Smolin (1)
- Lehre (1)
- Lernen (1)
- Lidar (1)
- Lifetime spectroscopy (1)
- Lightning (1)
- Link rate adaptation (1)
- Linkratenanpassung (1)
- Linux (1)
- LoRa (1)
- LoRaWAN (1)
- LoRaWan (1)
- Logic Programming (1)
- Logische Programmierung (1)
- Lunar Caves (1)
- Lunar Exploration (1)
- MAC (1)
- MDR (1)
- MTC (1)
- MVC <Software> (1)
- Mackenzie-River-Delta (1)
- Mapping (1)
- Mars (1)
- Mashup (1)
- Mashup <Internet> (1)
- Mathematische Modellierung (1)
- Mathematisches Modell (1)
- Measurement (1)
- Medical Image Analysis (1)
- Medium <Physik> (1)
- Medizin (1)
- Mehrebenensimulation (1)
- Mehrfahrzeugsysteme (1)
- Mehrkriterielle Optimierung (1)
- Mehrpfadübertragung (1)
- Mehrschichtnetze (1)
- Mehrschichtsystem (1)
- Mensch-Maschine-Kommunikation (1)
- Mesh Networks (1)
- Mesh Netze (1)
- Metaverse (1)
- Methode (1)
- Methodologie (1)
- Microarray (1)
- Middleware (1)
- Miniaturisierung (1)
- Minimally invasive vascular intervention (1)
- Missionsbetrieb (1)
- Mobile Roboter (1)
- Mobiles Internet (1)
- Mobilfunk (1)
- Modell (1)
- Modellbasierte Diagnose (1)
- Modellierungstechniken (1)
- Modelling (1)
- Modularität (1)
- Molecular systematics (1)
- Mond (1)
- Motion Planning (1)
- Multi-Agent-Simulation (1)
- Multi-Layer (1)
- Multi-Network Service (1)
- Multi-Netzwerk Dienste (1)
- Multi-Paradigm Programming (1)
- Multi-Paradigm Programming Framework (1)
- Multi-agent system (1)
- Multiagentensimulation (1)
- Multiagentensystem (1)
- Multipath Transmission (1)
- Multiple-Choice Examination (1)
- Multiple-Choice Prüfungen (1)
- Mycoplasma (1)
- Mycoplasma pneumoniae (1)
- NP (1)
- NP-Vollständigkeit (1)
- NP-complete sets (1)
- NP-hartes Problem (1)
- NP-schweres Problem (1)
- Naïve Bayesian (1)
- Network Emulator (1)
- Network Experiments (1)
- Network Management (1)
- Network Measurements (1)
- Network Virtualization (1)
- Networks (1)
- Netzplantechnik (1)
- Netzplanung (1)
- Netzvirtualisierung (1)
- Netzwerkplanung (1)
- Netzwerkvirtualisierung (1)
- Neuromuscular junctions (1)
- Neuronales Netz (1)
- Newton Methods (1)
- Newton-Verfahren (1)
- Next Generation Networks (1)
- Nichtholonome Fahrzeuge (1)
- Nichtlineare Regelung (1)
- Nuclear RDNA (1)
- Object Detection (1)
- Object-Oriented Programming (1)
- Objektorientierte Programmierung (1)
- Open Source (1)
- Operator (1)
- Optical Flow (1)
- Optimale Kontrolle (1)
- Optimierungsproblem (1)
- Optimization on Lie Groups (1)
- Overlapping (1)
- Overlay (1)
- Overlay Netzwerke (1)
- Overlay networks (1)
- Overlays (1)
- P4-INT (1)
- Panorama Images (1)
- Parameterkalibrierung (1)
- Partition <Mengenlehre> (1)
- Partitionen (1)
- Path Computation Element (1)
- Pattern Recognition (1)
- Peer-to-Peer (1)
- Performance Analysis (1)
- Performance Enhancing Proxies (1)
- Performance Management (1)
- Performance Modeling (1)
- Pfadberechnungselement (1)
- Picosatellite (1)
- Place of revelation (1)
- Planare Graphen (1)
- Planausführung (1)
- Planung (1)
- Planungssystem (1)
- Poisson surface reconstruction (1)
- PolSAR (1)
- Polyeder (1)
- Polypektomie (1)
- Positron annihilation spectroscopy (1)
- Post's Classes (1)
- Postsche Klassen (1)
- Prediction (1)
- Process Optimization (1)
- Processing Model (1)
- Processing model (1)
- Profile distances (1)
- Project Management (1)
- Projektmanagement (1)
- Prozessoptimierung (1)
- Publish-Subscribe-System (1)
- QUIC (1)
- QoS (1)
- Quality of Experience (QoE) (1)
- Quality of Experience QoE (1)
- Quality of Service (1)
- Quality of Service (QoS) (1)
- Quality-of-Experience (1)
- Quality-of-Service (1)
- Quality-of-Service (QoS) (1)
- Quantitative structure-property relationship modeling (QSPR) (1)
- Quantor (1)
- RBCL Gene-sequences (1)
- RGB-D (1)
- RNA sequencing (1)
- RRM (1)
- Radarfernerkundung (1)
- Raumdaten (1)
- Real-Time Operating Systems (1)
- Real-time (1)
- Rechenzentrum Universität Würzburg (1)
- Reconstruction of original text (1)
- Refactoring (1)
- Reference Architecture (1)
- Regelbasiertes System (1)
- Regelung (1)
- Registration (1)
- Registrierung (1)
- Registrierung <Bildverarbeitung> (1)
- Reguläre Sprache (1)
- Relief <Geografie> (1)
- Rendezvous (1)
- Reproducibility (1)
- Requirements Management (1)
- Resource and Performance Management (1)
- Ressourcen Management (1)
- Ressourcenallokation (1)
- Rettungsroboter (1)
- Risikomanagement (1)
- Risk Management (1)
- Robot (1)
- Roboter (1)
- Rule-based Systems (1)
- SARS-CoV-2 (1)
- SBA (1)
- SDN/NVF (1)
- SMLM (1)
- SNP (1)
- Scatter Plot (1)
- Scheduling (1)
- Search-and-Rescue (1)
- Secondary structure (1)
- Selbstkalibrierung (1)
- Self-Evaluation Programs (I2.399.780) (1)
- Self-calibration (1)
- Semantic Web (1)
- Semantics (1)
- Semantik (1)
- Sensing-aaS (1)
- Sensor (1)
- Septins (1)
- Service Mobility (1)
- Services (1)
- Sichtbarkeit (1)
- Similarity Measure (1)
- Simulator (1)
- Situationsbewusstsein (1)
- Skype (1)
- Small Satellites (1)
- Smart User Interaction (1)
- Social Web (1)
- Software (1)
- Software Performance Engineering (1)
- Software Performance Modeling (1)
- Software architecture (1)
- Software design (1)
- Software product lines (1)
- Softwareentwicklung (1)
- Source Code Visualization (1)
- Soziale Software (1)
- Spam-Mail (1)
- Spherical Robot (1)
- Spring (1)
- Stages of Prophet Mohammad’s messengership (1)
- Standardisierung (1)
- Standortproblem (1)
- Statistical classifiers (1)
- Statistics (1)
- Statistische Mechanik (1)
- Statistische Physik (1)
- Sternfreie Sprache (1)
- Steuerung (1)
- Stiffness (1)
- Stochastic Algorithms (1)
- Stochastik (1)
- Stochastikon (1)
- Stochastische Optimierung (1)
- Strahlentherapie (1)
- Straubing-Th´erien-Hierarchie (1)
- Straßennetzwerk (1)
- Straßenverkehr (1)
- Structure-from-Motion (1)
- Strukturelle Komplexität (1)
- Struts (1)
- Subgroup Mining (1)
- Subgruppenentdeckung (1)
- Substruktur (1)
- Suchverfahren (1)
- Support Vector Machine (1)
- Synapses (1)
- Synaptic vesicles (1)
- Synthetic Aperture Radar (1)
- System (1)
- Szenariogenerierung (1)
- TSN (1)
- TTL (1)
- TTL validation of data consistency (1)
- Teaching (1)
- Telematik (1)
- Terramechanics (1)
- Testbed (1)
- Text categorization (1)
- Text segmentation (1)
- Theoretical computer science (1)
- Thermografie (1)
- Time resolved measurements (1)
- Tomografie (1)
- Topografie (1)
- Torque (1)
- Trainingssystem (1)
- Travelling-salesman-Problem (1)
- Tumor motion (1)
- Tumorbewegung (1)
- U-Bahnlinienplan (1)
- UI and Interaction Design (1)
- UML Klassendiagramm (1)
- UML class diagram (1)
- UMTS (1)
- URL (1)
- URLLC (1)
- Underwater Mapping (1)
- Underwater Scanning (1)
- Unmanned Aerial Vehicle (1)
- Unstetige Regelung (1)
- Usability (1)
- Use case (1)
- User Behavior (1)
- User Participation (1)
- V-antigen (1)
- VNF (1)
- VPN (1)
- Variability (1)
- Venus (1)
- Veranstaltung (1)
- Verbotenes Muster (1)
- Verbände (1)
- Verkehrslenkung (1)
- Verteilung von Inhalten (1)
- Vesicles (1)
- Video Quality Monitoring (1)
- Video Streaming (1)
- Videoübertragung (1)
- Virtualisierung (1)
- Virtuelles Netzwerk (1)
- Visibility (1)
- Visual Text Mining (1)
- Visual Tracking (1)
- Visualized Kathará (1)
- Voice-over-IP (VoIP) (1)
- Volltextsuche (1)
- Vorhersage (1)
- WH2 domain (1)
- WLAN (1)
- WNT (1)
- Warteschlangentheorie (1)
- Web service (1)
- WebGL (1)
- Webmail-System (1)
- Webservice Composition (1)
- Werkstattdiagnose (1)
- WhatsApp (1)
- Wheel (1)
- Winkel (1)
- Wire relaxation (1)
- Wireless LAN (1)
- Wireless Sensor/Actuator Systems (1)
- Wissensbanksystem (1)
- Wissensbasiertes System (1)
- Wissenschaftliche Beobachtung (1)
- Wissensendeckung (1)
- Worterweiterungen (1)
- Wrapper (1)
- Wrappers (1)
- XML model (1)
- XR-artificial intelligence combination (1)
- XR-artificial intelligence continuum (1)
- Yersinia enterocolitica (1)
- Yolk protein (1)
- YouTube (1)
- Zebrafish (1)
- Zeichnen von Graphen (1)
- Zeitdiskretes System (1)
- Zugangskontrolle (1)
- Zählprobleme (1)
- abgeschlossene Klassen (1)
- acrophobia (1)
- actin nucleation (1)
- adaptation (1)
- adaptation models (1)
- administrative boundary (1)
- admission control (1)
- adult learning (1)
- advertising effectiveness (1)
- aerodynamics (1)
- aerospace (1)
- aftermarket diagnostic (1)
- agent-based models (1)
- agents (1)
- agile Prozesse (1)
- agile processes (1)
- aging (1)
- alignment (1)
- anamnesis tool (1)
- aneurysm (1)
- angular schematization (1)
- annotation (1)
- anomaly detection (1)
- anomaly prediction (1)
- ant-colony optimization (1)
- anthropomorphism (1)
- anxiety (1)
- apixaban (1)
- application design (1)
- approximation algorithms (1)
- arabidopsis thaliana (1)
- arabidpsis thaliana (1)
- architectural design (1)
- arithmetic calculations (1)
- artificial intelligence education (1)
- artificial intelligence literacy (1)
- augmentation (1)
- automatic Layout (1)
- automatisches Layout (1)
- autonomic orchestration (1)
- autonomous (1)
- autonomous UAV (1)
- autorotation (1)
- availability (1)
- avatar embodiment (1)
- avatars (1)
- background knowledge (1)
- baseline detection (1)
- behavior (1)
- behavior change (1)
- behavior perception (1)
- bibliometric analysis (1)
- binary decision diagram (1)
- binary tanglegram (1)
- biofuel (1)
- biohybrid systems (1)
- bioinformatics (1)
- biological development (1)
- biomanufacturing (1)
- biosignals (1)
- bit (1)
- bit-parallel (1)
- boundary labeling (1)
- brain (1)
- building (1)
- caenorhabditis elegans (1)
- camera orientation (1)
- car-like robots (1)
- carbon (1)
- carboxylation (1)
- cardiac magnetic resonance (1)
- case study (1)
- caspase-3 (1)
- cell membranes (1)
- cerebral ischemia (1)
- certifying algorithm (1)
- chain cover (1)
- channel management (1)
- cisplatin (1)
- classification (1)
- climate (1)
- cloud-native (1)
- co-authorships (1)
- co-inventorships (1)
- cognitive impairment (1)
- coherence (1)
- collaboration (1)
- collision (1)
- colony-stimulating factor (1)
- combination therapy (1)
- communication models (1)
- communication networks (1)
- community detection (1)
- comparative sequence analysis (1)
- competitive location (1)
- complex traits (1)
- complexity (1)
- compressed sensing (1)
- computational (1)
- computational complexity (1)
- computer performance evaluation (1)
- computer virus (1)
- computergestützte Softwaretechnik (1)
- computers as social actors (1)
- concept search (1)
- condition prediction (1)
- congruence (1)
- connector (1)
- constrained forest (1)
- contact representation (1)
- container virtualization (1)
- content-based image retrieval (1)
- continuous-time SLAM (1)
- conversational agent (1)
- conversational agents (1)
- convex bipartite graph (1)
- convolutional neural network (1)
- corticotropin-releasing hormone (1)
- cost-sensitive learning (1)
- counting problems (1)
- crosstalk (1)
- crowdsourced QoE measurements (1)
- crowdsourced measurements (1)
- crowdsourced network measurements (1)
- cryo-EM (1)
- cryo-ET (1)
- crystal growth (1)
- crystallization (1)
- cultural and media studies (1)
- culturally aware (1)
- cuneiform (1)
- curves (1)
- cyber-physical systems (1)
- cybersickness (1)
- cytokine profiling (1)
- d3web.Train (1)
- dSTORM (1)
- data mining (1)
- data plane programming (1)
- data stream processing (1)
- data structure (1)
- data warehouse (1)
- dataplane programming (1)
- decision support system (1)
- decision-making (1)
- decission finding (1)
- decoding error rate (1)
- deep metric learning (1)
- definite clause grammars (1)
- deformation-based method (1)
- delay QoS exponent (1)
- delay bound violation probability (1)
- delay constrained (1)
- dementia (1)
- descent (1)
- design cycle (1)
- detection time simulation (1)
- diagnostic accuracy (1)
- dial a ride (1)
- differentiation (1)
- digital twin (1)
- dimensions of proximity (1)
- direct oral anticoagulants (1)
- direct thrombin inhibitor (1)
- discrete-time analysis (1)
- disease (1)
- disjoint multi-paths (1)
- disruption project (1)
- distance-based classifier (1)
- distributed control (1)
- docker (1)
- driving simulation (1)
- drug (1)
- drug-minded protein (1)
- dynamic flow migration (1)
- dynamic programming (1)
- dynamische Umgebungen (1)
- eHealth (1)
- early printed books (1)
- eco-metabolomics (1)
- ecological momentary assessment (1)
- edge labeled graphs (1)
- effective Bandwidth (1)
- efficient algorithm (1)
- electroencephalography (1)
- electrolytes (1)
- electronic health records (1)
- elementary mode analysis (1)
- elementary modes (1)
- elevated plus-maze (1)
- embedding techniques (1)
- emergent time (1)
- emotions (1)
- empathy (1)
- emulation (1)
- encryption (1)
- end user (1)
- endurance (1)
- energy efficiency (1)
- environmental sound (1)
- enzyme (1)
- event detection (1)
- event-related potentials-ERP (1)
- exercise intensity (1)
- experience (1)
- experimental evaluation (1)
- expertise framing (Min5-Max 8) (1)
- expression (1)
- expression signature (1)
- extended reality (1)
- extended reality (XR) (1)
- factor XA inhibitor (1)
- failure prediction (1)
- fast reroute (1)
- fault detection (1)
- feature matching (1)
- feature-matching (1)
- federated learning (1)
- few-shot learning (1)
- field-programmable architectures (1)
- field-programmable gate arrays (1)
- firewall (1)
- fixed-parameter tractability (1)
- flies (1)
- fluoroscopy (1)
- food quality (1)
- force dynamics (1)
- foreign language learning and teaching (1)
- formation driving (1)
- formation flight (1)
- fruit temperature (1)
- full-text search (1)
- functional analysis (1)
- future Internet architecture (1)
- future energy grid exploration (1)
- game mechanics (1)
- gamification (1)
- gamma (1)
- generative systems (1)
- genes (1)
- genetic algorithm (1)
- genetic regulatory network (1)
- geospatial data (1)
- global IPX network (1)
- graph (1)
- graph algorithm (1)
- graph decomposition (1)
- graphs (1)
- green systems biology (1)
- group-based communication (1)
- handwriting (1)
- handwritten artefact recognition (1)
- haptic data (1)
- hardness (1)
- hardware-in-the-loop simulation (1)
- hardware-in-the-loop streaming system (1)
- hepatotoxicity (1)
- heuristics (1)
- hierarchy (1)
- histidine kinase (1)
- historical images (1)
- hit ratio analysis and simulation (1)
- homology modeling (1)
- hospital data (1)
- human body weight (1)
- human computer interaction (HCI) (1)
- human-artificial intelligence interaction (1)
- human-artificial intelligence interface (1)
- human-centered design (1)
- human-centered, human-robot (1)
- humantechnology interaction (1)
- human–computer interaction (1)
- hybrid Diagnostic (1)
- hybrid access (1)
- hybride Diagnose (1)
- hypotonic (1)
- hypotonic solutions (1)
- illusion of self-motion (1)
- image classification (1)
- image processing (1)
- image schemas (1)
- imbalanced regression (1)
- immersive advertising (1)
- immersive classroom (1)
- immersive classroom management (1)
- immersive learning technologies (1)
- immunity (1)
- implicit association test (1)
- in situ analysis (1)
- independent crossing (1)
- induced matching (1)
- inflation (1)
- informal education (1)
- information extraction (1)
- information systems and information technology (1)
- inhibitor (1)
- intelligent transportation systems (1)
- intelligent vehicles (1)
- intelligent virtual agents (1)
- intelligent voice assistant (1)
- intelligente Applikationen (1)
- intention-behavior-gap (1)
- inter-coder reliability (1)
- interaction (1)
- interactive authoring system (1)
- interactive collation of textual variants (1)
- intercultural learning and teaching (1)
- interdisciplinary education (1)
- intermediate host (1)
- internal transcribed spacer 2 (1)
- internet protocol (1)
- internet traffic (1)
- interpolation (1)
- intervention design (1)
- intervention evaluation (1)
- intraoperative imaging (1)
- invasive vascular interventions (1)
- iowa gambling task (1)
- isotonic (1)
- key-insight extraction (1)
- kinect (1)
- labeling (1)
- land-cover area (1)
- landing (1)
- language-image pre-training (1)
- latency (1)
- lattices (1)
- layout recognition (1)
- learning environments (1)
- least cost (1)
- life-span regulation (1)
- lifetime spectroscopy (1)
- light-gated proteins (1)
- load balancing (1)
- local energy system (1)
- locomotion (1)
- logic programming (1)
- logistics (1)
- long-term analysis (1)
- lymphotoxicity (1)
- malaria (1)
- map projections (1)
- mapping (1)
- markers (1)
- mathematical model (1)
- measurement (1)
- media analysis (1)
- media equation (1)
- medical analytics (1)
- medical device regulation (1)
- medical device software (1)
- medical records (1)
- medieval manuscripts (1)
- meditation (1)
- membrane protein (1)
- membrane proteins (1)
- memory immune responses (1)
- metabolic flux (1)
- metabolism (1)
- metabolomics (1)
- metastasis (1)
- methylene blue (1)
- metro map (1)
- mice (1)
- microbes (1)
- mindfulness (1)
- misconceptions (1)
- mission operation (1)
- mixed reality (1)
- mixed-cultural (1)
- mixed-cultural settings (1)
- mobile application (1)
- mobile instant messaging (1)
- mobile messaging application (1)
- mobile streaming (1)
- model following (1)
- model output statistics (1)
- model predictive control (1)
- model-base diagnosis (1)
- model-based diagnosis (1)
- modeling techniques (1)
- modules (1)
- molecular systematics (1)
- monotone drawing (1)
- morphing (1)
- mouse (1)
- multi-vehicle formations (1)
- multi-vehicle rendezvous (1)
- multimodal fusion (1)
- multimodal interface (1)
- multimodal learning (1)
- multipath communication (1)
- multipath packet scheduling (1)
- multiple myeloma (1)
- multirotors (1)
- multiscale encoder (1)
- mutation (1)
- mycoplasma (1)
- n-Gramm (1)
- n-gram (1)
- nano-satellite (1)
- nanocellulose (1)
- natural interfaces (1)
- natural language processing (1)
- natural language processing · · · (1)
- natural user interfaces (1)
- network (1)
- network design (1)
- network planning (1)
- network softwarization (1)
- network upgrade (1)
- network virtualization (1)
- networked robotics (1)
- networking (1)
- neume notation (1)
- neural architecture (1)
- noise measurement (1)
- non-native accent (1)
- non-terrestrial networks (1)
- nonholonomic vehicles (1)
- nonhuman-primates (1)
- nonverbal behavior (1)
- object detection (1)
- octree (1)
- omics (1)
- open source (1)
- optical music recognition (1)
- optimization (1)
- orchestration (1)
- organogenesis (1)
- origin (1)
- overprovisioning (1)
- oxidative stress (1)
- packet reception method (1)
- painful (1)
- pangolin (1)
- particle picking (1)
- partitions (1)
- passage of time (1)
- passive haptic feedback (1)
- path computation (1)
- pathway (1)
- pattern perception (1)
- performance (1)
- performance analysis (1)
- performance evaluation (1)
- performance liquid chromatography (1)
- performance monitoring (1)
- performance prediction (1)
- permeability (1)
- pestis infection (1)
- phase space (1)
- phase transition (1)
- photorespiration (1)
- phylogenetic tree (1)
- phylogeny (1)
- place-illusion (1)
- plain orchestrating service (1)
- plan execution (1)
- plausibility (1)
- plausibility-illusion (1)
- pneumoniae (1)
- pneumonic plague (1)
- point cloud (1)
- point cloud compression (1)
- point-to-plane measure (1)
- point-to-point measure (1)
- pollution (1)
- pos (1)
- posets (1)
- positioning (1)
- precision horticulture (1)
- precision training (1)
- presence (1)
- private chat groups (1)
- procedural content generation (1)
- procedural fusion methods (1)
- process model (1)
- processing pipeline (1)
- progeria (1)
- promoter (1)
- prompt engineering (1)
- protein (1)
- protein chip (1)
- protein-interaction networks (1)
- pseudomas-syringae (1)
- psychomotor training (1)
- psychophyisology (1)
- public speaking (1)
- pulse simulation (1)
- q-Gramm (1)
- q-gram (1)
- quadcopter (1)
- quadcopters (1)
- quality assurance (1)
- quality evaluation (1)
- quality of experience (1)
- quality of experience prediction (1)
- quantification (1)
- qubit (1)
- radio resource management (1)
- radiology (1)
- ransomware (1)
- real world evidence (1)
- real-world application (1)
- realism (1)
- receding horizon control (1)
- receptor (1)
- recombinant protein rVE (1)
- recombination (1)
- recommender system (1)
- reconfiguration (1)
- regelbasierte Nachbearbeitung (1)
- reload cost (1)
- remote control (1)
- research methods (1)
- resilience (1)
- response regulator (1)
- ribosomal RNA (1)
- rich vehicle routing problem (1)
- richtersius coronifer (1)
- right angle crossing (1)
- road network (1)
- robotics (1)
- robustness (1)
- rotorcraft (1)
- rotors (1)
- routing (1)
- rule based post processing (1)
- sample weighting (1)
- satisfiability problems (1)
- scalability evaluation (1)
- scalable quadcopter (1)
- scenario creation (1)
- science, technology and society (1)
- secondary structure (1)
- secure group communication (1)
- self-adaptive (1)
- self-adaptive systems (1)
- self-assembly (1)
- self-aware computing systems (1)
- self-managing systems (1)
- semantic fusion (1)
- semantic understanding (1)
- semantic web (1)
- semantical aesthetic (1)
- semantische Ästhetik (1)
- sensitivity analysis (1)
- sensor devices (1)
- sensor fusion (1)
- sensor network (1)
- sensor networks (1)
- sentinel (1)
- sequence alignment (1)
- serious games (1)
- serum (1)
- service based software architecture (1)
- service brokerage (1)
- service-curve estimation (1)
- sesnsors (1)
- set (1)
- shootin-1 (1)
- short block-length (1)
- shortest path routing (1)
- signal processing (1)
- signaling traffic (1)
- simulation system (1)
- simultaneous embedding (1)
- single-electron transistors (1)
- skalierbare Diagnose (1)
- sketching (1)
- slam (1)
- smart meter data utilization (1)
- smooth orthogonal drawing (1)
- snow shoveling (1)
- social VR (1)
- social interaction (1)
- social relationship (1)
- social robot (1)
- social robotics (1)
- social role (1)
- socially interactive agents (1)
- software defined network (1)
- spam mail (1)
- spanning tree (1)
- spire (1)
- stability (1)
- stable state (1)
- standardization (1)
- state management (1)
- statistical validity (1)
- statistics and numerical data (1)
- stereotypes (1)
- stochastic thinking (1)
- stochastisches Denken (1)
- stream processing (1)
- stroke (1)
- structural complexity (1)
- student simulation (1)
- stylus (1)
- sun exposure (1)
- sunburn (1)
- superoxide-dismutase (1)
- supervised learning (1)
- surface model (1)
- survey (1)
- survival (1)
- sustainability (1)
- switching navigation (1)
- synthetic biology (1)
- synthetic pathways (1)
- system (1)
- system architecture design (1)
- systematic literature review (1)
- systematic review (1)
- table extraction (1)
- table understanding (1)
- taxonomy (1)
- teacher education (1)
- technology-supported learning (1)
- telematics (1)
- temperature (1)
- text categorization (1)
- text line detection (1)
- text supervision (1)
- theory (1)
- therapeutic application (1)
- thermal camera (1)
- thermal point cloud (1)
- time calibration (1)
- time perception (1)
- time series (1)
- timestamping method (1)
- tolerance (1)
- tomography (1)
- tonicity (1)
- tools (1)
- topology (1)
- training systems (1)
- trait anxiety (1)
- trajectory planning (1)
- transcription (1)
- transformations (1)
- transformer (1)
- translational neuroscience (1)
- transmission (1)
- transport microenvironments (1)
- transport protocols (1)
- transportation (1)
- tree (1)
- trust (1)
- trustworthiness (1)
- university network (1)
- unmanned aerial vehicle (1)
- unmanned aerial vehicles (1)
- usability evaluation (1)
- use cases (1)
- user experience (1)
- user interaction (1)
- user interfaces (1)
- user study (1)
- user-generated content (1)
- v (1)
- vaccine (1)
- validation (1)
- vection (1)
- vehicle dynamics (1)
- vehicular navigation (1)
- verbal behaviour (1)
- vernetzte Roboter (1)
- virtual agent interaction (1)
- virtual audience (1)
- virtual humans (1)
- virtual queue (1)
- virtual reality training (1)
- virtual stimuli (1)
- virtual tunnel (1)
- virtual-reality-continuum (1)
- visual analytics (1)
- visual proteomics (1)
- vitellogenin (1)
- voice assistant (1)
- voice-based artificial intelligence (1)
- vom Nutzer erfahrene Dienstgüte QoE (1)
- voting location (1)
- water stress (1)
- waypoint parameter (1)
- wearable (1)
- webmail system (1)
- wheel (1)
- wireless network (1)
- word clouds (1)
- word extensions (1)
- zooming (1)
- zukünftige Kommunikationsnetze (1)
- zukünftiges Internet (1)
- Ähnlichkeitsmaß (1)
- Überlappung (1)
Institute
- Institut für Informatik (203)
- Theodor-Boveri-Institut für Biowissenschaften (29)
- Institut Mensch - Computer - Medien (17)
- Institut für deutsche Philologie (17)
- Institut für Klinische Epidemiologie und Biometrie (7)
- Rechenzentrum (7)
- Center for Computational and Theoretical Biology (4)
- Graduate School of Science and Technology (3)
- Medizinische Klinik und Poliklinik II (3)
- Institut für Funktionsmaterialien und Biofabrikation (2)
Schriftenreihe
Sonstige beteiligte Institutionen
- Cologne Game Lab (2)
- Birmingham City University (1)
- DATE Lab, KITE Research Insititute, University Health Network, Toronto, Canada (1)
- EMBL Heidelberg (1)
- INAF Padova, Italy (1)
- Jacobs University Bremen, Germany (1)
- Open University of the Netherlands (1)
- Servicezentrum Medizin-Informatik (Universitätsklinikum) (1)
- Social and Technological Systems (SaTS) lab, School of Art, Media, Performance and Design, York University, Toronto, Canada (1)
- TH Köln (1)
Dynamic point cloud compression based on projections, surface reconstruction and video compression
(2021)
In this paper we will present a new dynamic point cloud compression based on different projection types and bit depth, combined with the surface reconstruction algorithm and video compression for obtained geometry and texture maps. Texture maps have been compressed after creating Voronoi diagrams. Used video compression is specific for geometry (FFV1) and texture (H.265/HEVC). Decompressed point clouds are reconstructed using a Poisson surface reconstruction algorithm. Comparison with the original point clouds was performed using point-to-point and point-to-plane measures. Comprehensive experiments show better performance for some projection maps: cylindrical, Miller and Mercator projections.
A binary tanglegram is a drawing of a pair of rooted binary trees whose leaf sets are in one-to-one correspondence; matching leaves are connected by inter-tree edges. For applications, for example, in phylogenetics, it is essential that both trees are drawn without edge crossings and that the inter-tree edges have as few crossings as possible. It is known that finding a tanglegram with the minimum number of crossings is NP-hard and that the problem is fixed-parameter tractable with respect to that number.
We prove that under the Unique Games Conjecture there is no constant-factor approximation for binary trees. We show that the problem is NP-hard even if both trees are complete binary trees. For this case we give an O(n 3)-time 2-approximation and a new, simple fixed-parameter algorithm. We show that the maximization version of the dual problem for binary trees can be reduced to a version of MaxCut for which the algorithm of Goemans and Williamson yields a 0.878-approximation.
Webservices composition is traditionally carried out using composition technologies such as Business Process Execution Language (BPEL) [1] and Web Service Choreography Interface (WSCI) [2]. The composition technology involves the process of web service discovery, invocation, and composition. However these technologies are not easy and flexible enough because they are mainly developer-centric. Moreover majority of websites have not yet embarked into the world of web service, although they have very important and useful information to offer. Is it because they have not understood the usefulness of web services or is it because of the costs? Whatever might be the answers to these questions, time and money are definitely required in order to create and offer web services. To avoid these expenditures, wrappers [7] to automatically generate webservices from websites would be a cheaper and easier solution. Mashups offer a different way of doing webservices composition. In web environment a Mashup is a web application that brings together data from several sources using webservices, APIs, wrappers and so on, in order to create entirely a new application that was not provided before. This paper presents first an overview of Mashups and the process of web service invocation and composition based on Mashup, then describes an example of a web-based simulator for navigation system in Germany.
3D point clouds are a de facto standard for 3D documentation and modelling. The advances in laser scanning technology broadens the usability and access to 3D measurement systems. 3D point clouds are used in many disciplines such as robotics, 3D modelling, archeology and surveying. Scanners are able to acquire up to a million of points per second to represent the environment with a dense point cloud. This represents the captured environment with a very high degree of detail. The combination of laser scanning technology with photography adds color information to the point clouds. Thus the environment is represented more realistically. Full 3D models of environments, without any occlusion, require multiple scans. Merging point clouds is a challenging process. This thesis presents methods for point cloud registration based on the panorama images generated from the scans. Image representation of point clouds introduces 2D image processing methods to 3D point clouds. Several projection methods for the generation of panorama maps of point clouds are presented in this thesis. Additionally, methods for point cloud reduction and compression based on the panorama maps are proposed. Due to the large amounts of data generated from the 3D measurement systems these methods are necessary to improve the point cloud processing, transmission and archiving. This thesis introduces point cloud processing methods as a novel framework for the digitisation of archeological excavations. The framework replaces the conventional documentation methods for excavation sites. It employs point clouds for the generation of the digital documentation of an excavation with the help of an archeologist on-site. The 3D point cloud is used not only for data representation but also for analysis and knowledge generation. Finally, this thesis presents an autonomous indoor mobile mapping system. The mapping system focuses on the sensor placement planning method. Capturing a complete environment requires several scans. The sensor placement planning method solves for the minimum required scans to digitise large environments. Combining this method with a navigation system on a mobile robot platform enables it to acquire data fully autonomously. This thesis introduces a novel hole detection method for point clouds to detect obscured parts of a captured environment. The sensor placement planning method selects the next scan position with the most coverage of the obscured environment. This reduces the required number of scans. The navigation system on the robot platform consist of path planning, path following and obstacle avoidance. This guarantees the safe navigation of the mobile robot platform between the scan positions. The sensor placement planning method is designed as a stand alone process that could be used with a mobile robot platform for autonomous mapping of an environment or as an assistant tool for the surveyor on scanning projects.
DLTPulseGenerator: a library for the simulation of lifetime spectra based on detector-output pulses
(2018)
The quantitative analysis of lifetime spectra relevant in both life and materials sciences presents one of the ill-posed inverse problems and, hence, leads to most stringent requirements on the hardware specifications and the analysis algorithms. Here we present DLTPulseGenerator, a library written in native C++ 11, which provides a simulation of lifetime spectra according to the measurement setup. The simulation is based on pairs of non-TTL detector output-pulses. Those pulses require the Constant Fraction Principle (CFD) for the determination of the exact timing signal and, thus, the calculation of the time difference i.e. the lifetime. To verify the functionality, simulation results were compared to experimentally obtained data using Positron Annihilation Lifetime Spectroscopy (PALS) on pure tin.
Two-component systems (TCS) are short signalling pathways generally occurring in prokaryotes. They frequently regulate prokaryotic stimulus responses and thus are also of interest for engineering in biotechnology and synthetic biology. The aim of this study is to better understand and describe rewiring of TCS while investigating different evolutionary scenarios. Based on large-scale screens of TCS in different organisms, this study gives detailed data, concrete alignments, and structure analysis on three general modification scenarios, where TCS were rewired for new responses and functions: (i) exchanges in the sequence within single TCS domains, (ii) exchange of whole TCS domains; (iii) addition of new components modulating TCS function. As a result, the replacement of stimulus and promotor cassettes to rewire TCS is well defined exploiting the alignments given here. The diverged TCS examples are non-trivial and the design is challenging. Designed connector proteins may also be useful to modify TCS in selected cases.
Das stochastische Denken, die Bernoullische Stochastik und dessen informationstechnologische Umsetzung, namens Stochastikon stellen die Grundlage für das Verständnis und die erfolgreiche Nutzung einer stochastischen Wissenschaft dar. Im Rahmen dieser Arbeit erfolgt eine Klärung des Begriffs des stochastischen Denkens, eine anschauliche Darstellung der von Elart von Collani entwickelten Bernoullischen Stochastik und eine Beschreibung von Stochastikon. Dabei werden sowohl das Gesamtkonzept von Stochastikon, sowie die Ziele, Aufgaben und die Realisierung der beiden Teilsysteme namens Mentor und Encyclopedia vorgestellt. Das stochastische Denken erlaubt eine realitätsnahe Sichtweise der Dinge, d.h. eine Sichtweise, die mit den menschlichen Beobachtungen und Erfahrungen im Einklang steht und somit die Unsicherheit über zukünftige Entwicklungen berücksichtigt. Der in diesem Kontext verwendete Begriff der Unsicherheit bezieht sich ausschließlich auf zukünftige Entwicklungen und äußert sich in Variabilität. Quellen der Unsicherheit sind einerseits die menschliche Ignoranz und andererseits der Zufall. Unter Ignoranz wird hierbei die Unwissenheit des Menschen über die unbekannten, aber feststehenden Fakten verstanden, die die Anfangsbedingungen der zukünftigen Entwicklung repräsentieren. Die Bernoullische Stochastik liefert ein Regelwerk und ermöglicht die Entwicklung eines quantitativen Modells zur Beschreibung der Unsicherheit und expliziter Einbeziehung der beiden Quellen Ignoranz und Zufall. Das Modell trägt den Namen Bernoulli-Raum und bildet die Grundlage für die Herleitung quantitativer Verfahren, um zuverlässige und genaue Aussagen sowohl über die nicht-existente zufällige Zukunft (Vorhersageverfahren), als auch über die unbekannte feststehende Vergangenheit (Messverfahren). Das Softwaresystem Stochastikon implementiert die Bernoullische Stochastik in Form einer Reihe autarker, miteinander kommunizierender Teilsysteme. Ziel des Teilsystems Encyclopedia ist die Bereitstellung und Bewertung stochastischen Wissens. Das Teilsystem Mentor dient der Unterstützung des Anwenders bei der Problemlösungsfindung durch Identifikation eines richtigen Modells bzw. eines korrekten Bernoulli-Raums. Der Lösungsfindungsprozess selber enthält keinerlei Unsicherheit. Die ganze Unsicherheit steckt in der Lösung, d.h. im Bernoulli-Raum, der explizit die vorhandene Unwissenheit (Ignoranz) und den vorliegenden Zufall abdeckend enthält.
The concept of digital literacy has been introduced as a new cultural technique, which is regarded as essential for successful participation in a (future) digitized world. Regarding the increasing importance of AI, literacy concepts need to be extended to account for AI-related specifics. The easy handling of the systems results in increased usage, contrasting limited conceptualizations (e.g., imagination of future importance) and competencies (e.g., knowledge about functional principles). In reference to voice-based conversational agents as a concrete application of AI, the present paper aims for the development of a measurement to assess the conceptualizations and competencies about conversational agents. In a first step, a theoretical framework of “AI literacy” is transferred to the context of conversational agent literacy. Second, the “conversational agent literacy scale” (short CALS) is developed, constituting the first attempt to measure interindividual differences in the “(il) literate” usage of conversational agents. 29 items were derived, of which 170 participants answered. An explanatory factor analysis identified five factors leading to five subscales to assess CAL: storage and transfer of the smart speaker’s data input; smart speaker’s functional principles; smart speaker’s intelligent functions, learning abilities; smart speaker’s reach and potential; smart speaker’s technological (surrounding) infrastructure. Preliminary insights into construct validity and reliability of CALS showed satisfying results. Third, using the newly developed instrument, a student sample’s CAL was assessed, revealing intermediated values. Remarkably, owning a smart speaker did not lead to higher CAL scores, confirming our basic assumption that usage of systems does not guarantee enlightened conceptualizations and competencies. In sum, the paper contributes to the first insights into the operationalization and understanding of CAL as a specific subdomain of AI-related competencies.
Das Potenzial der Wissensentdeckung in Daten wird häufig nicht ausgenutzt, was hauptsächlich auf Barrieren zwischen dem Entwicklerteam und dem Endnutzer des Data-Mining zurückzuführen ist. In dieser Arbeit wird ein transparenter Ansatz zum Beschreiben und Erklären von Daten für Entscheidungsträger vorgestellt. In Entscheidungsträger-zentrierten Aufgaben werden die Projektanforderungen definiert und die Ergebnisse zu einer Geschichte zusammengestellt. Eine Anforderung besteht dabei aus einem tabellarischen Bericht und ggf. Mustern in seinem Inhalt, jeweils verständlich für einen Entscheidungsträger. Die technischen Aufgaben bestehen aus einer Datenprüfung, der Integration der Daten in einem Data-Warehouse sowie dem Generieren von Berichten und dem Entdecken von Mustern wie in den Anforderungen beschrieben. Mehrere Data-Mining-Projekte können durch Wissensmanagement sowie eine geeignete Infrastruktur voneinander profitieren. Der Ansatz wurde in zwei Projekten unter Verwendung von ausschließlich Open-Source-Software angewendet.
Design and Implementation of Architectures for Interactive Textual Documents Collation Systems
(2011)
One of the main purposes of textual documents collation is to identify a base text or closest witness to the base text, by analyzing and interpreting differences also known as types of changes that might exist between those documents. Based on this fact, it is reasonable to argue that, explicit identification of types of changes such as deletions, additions, transpositions, and mutations should be part of the collation process. The identification could be carried out by an interpretation module after alignment has taken place. Unfortunately existing collation software such as CollateX1 and Juxta2’s collation engine do not have interpretation modules. In fact they implement the Gothenburg model [1] for collation process which does not include an interpretation unit. Currently both CollateX and Juxta’s collation engine do not distinguish in their critical apparatus between the types of changes, and do not offer statistics about those changes. This paper presents a model for both integrated and distributed collation processes that improves the Gothenburg model. The model introduces an interpretation component for computing and distinguishing between the types of changes that documents could have undergone. Moreover two architectures implementing the model in order to solve the problem of interactive collation are discussed as well. Each architecture uses CollateX library, and provides on the one hand preprocessing functions for transforming input documents into CollateX input format, and on the other hand a post-processing module for enabling interactive collation. Finally simple algorithms for distinguishing between types of changes, and linking collated source documents with the collation results are also introduced.
Learning a book in general involves reading it, underlining important words, adding comments, summarizing some passages, and marking up some text or concepts. Once deeper understanding is achieved, one would like to organize and manage her/his knowledge in such a way that, it could be easily remembered and efficiently transmitted to others. This paper discusses about modeling religious texts using semantic XML markup based on frame-based knowledge representation, with the purpose of assisting understanding, retention, and sharing of knowledge they contain. In this study, books organized in terms of chapters made up of verses are considered as the source of knowledge to model. Some metadata representing the multiple perspectives of knowledge modeling are assigned to each chapter and verse. Chapters and verses with their metadata form a meta-model, which is represented using frames, and published on a web mashup. An XML-based annotation and visualization system equipped with user interfaces for creating static and dynamic metadata, annotating chapters’ contents according to user selected semantics, and templates for publishing generated knowledge on the Internet, has been developed. The system has been applied to the Quran, and the result obtained shows that multiple perspectives of information modeling can be successfully applied to religious texts, in order to support analysis, understanding, and retention of the texts.
Today’s Internet architecture was not designed from scratch but was driven by new services that emerged during its development. Hence, it is often described as patchwork where additional patches are applied in case new services require modifications to the existing architecture. This process however is rather slow and hinders the development of innovative network services with certain architecture or network requirements. Currently discussed technologies like Software-Defined Networking (SDN) or Network Virtualization (NV) are seen as key enabling technologies to overcome this rigid best effort legacy of the Internet. Both technologies offer the possibility to create virtual networks that accommodate the specific needs of certain services. These logical networks are operated on top of a physical substrate and facilitate flexible network resource allocation as physical resources can be added and removed depending on the current network and load situation. In addition, the clear separation and isolation of networks foster the development of application-aware networks that fulfill the special requirements of emerging applications. A prominent use case that benefits from these extended capabilities of the network is denoted with service component mobility. Services hosted on Virtual Machines (VMs) follow their consuming mobile endpoints, so that access latency as well as consumed network resources are reduced. Especially for applications like video streaming, which consume a large fraction of the available resources, is this an important means to relieve the resource constraints and eventually provide better service quality. Service and endpoint mobility both allow an adaptation of the used paths between an offered service, i.e., video streaming and the consuming users in case the service quality drops due to network problems. To make evidence-based adaptations in case of quality drops, a scalable monitoring component is required that is able to monitor the service quality for video streaming applications with reliable accuracy. This monograph details challenges that arise when deploying a certain service, i.e., video streaming, in a future virtualized network architecture and discusses possible solutions. In particular, this work evaluates the performance of mechanisms enabling service mobility and presents an optimized architecture for service mobility. Concerning endpoint mobility, improvements are developed that reduce the latency between endpoints and consumed services and ensure connectivity regardless of the used mobile access network. In the last part, a network-based video quality monitoring solution is developed and its accuracy is evaluated.
In many real world settings, imbalanced data impedes model performance of learning algorithms, like neural networks, mostly for rare cases. This is especially problematic for tasks focusing on these rare occurrences. For example, when estimating precipitation, extreme rainfall events are scarce but important considering their potential consequences. While there are numerous well studied solutions for classification settings, most of them cannot be applied to regression easily. Of the few solutions for regression tasks, barely any have explored cost-sensitive learning which is known to have advantages compared to sampling-based methods in classification tasks. In this work, we propose a sample weighting approach for imbalanced regression datasets called DenseWeight and a cost-sensitive learning approach for neural network regression with imbalanced data called DenseLoss based on our weighting scheme. DenseWeight weights data points according to their target value rarities through kernel density estimation (KDE). DenseLoss adjusts each data point’s influence on the loss according to DenseWeight, giving rare data points more influence on model training compared to common data points. We show on multiple differently distributed datasets that DenseLoss significantly improves model performance for rare data points through its density-based weighting scheme. Additionally, we compare DenseLoss to the state-of-the-art method SMOGN, finding that our method mostly yields better performance. Our approach provides more control over model training as it enables us to actively decide on the trade-off between focusing on common or rare cases through a single hyperparameter, allowing the training of better models for rare data points.
To deliver the best user experience (UX), the human-centered design cycle (HCDC) serves as a well-established guideline to application developers. However, it does not yet cover network-specific requirements, which become increasingly crucial, as most applications deliver experience over the Internet. The missing network-centric view is provided by Quality of Experience (QoE), which could team up with UX towards an improved overall experience. By considering QoE aspects during the development process, it can be achieved that applications become network-aware by design. In this paper, the Quality of Experience Centered Design Cycle (QoE-CDC) is proposed, which provides guidelines on how to design applications with respect to network-specific requirements and QoE. Its practical value is showcased for popular application types and validated by outlining the design of a new smartphone application. We show that combining HCDC and QoE-CDC will result in an application design, which reaches a high UX and avoids QoE degradation.
The landscape of today’s programming languages is manifold. With the diversity of applications, the difficulty of adequately addressing and specifying the used programs increases. This often leads to newly designed and implemented domain-specific languages. They enable domain experts to express knowledge in their preferred format, resulting in more readable and concise programs. Due to its flexible and declarative syntax without reserved keywords, the logic programming language Prolog is particularly suitable for defining and embedding domain-specific languages.
This thesis addresses the questions and challenges that arise when integrating domain-specific languages into Prolog. We compare the two approaches to define them either externally or internally, and provide assisting tools for each. The grammar of a formal language is usually defined in the extended Backus–Naur form. In this work, we handle this formalism as a domain-specific language in Prolog, and define term expansions that allow to translate it into equivalent definite clause grammars. We present the package library(dcg4pt) for SWI-Prolog, which enriches them by an additional argument to automatically process the term’s corresponding parse tree. To simplify the work with definite clause grammars, we visualise their application by a web-based tracer.
The external integration of domain-specific languages requires the programmer to keep the grammar, parser, and interpreter in sync. In many cases, domain-specific languages can instead be directly embedded into Prolog by providing appropriate operator definitions. In addition, we propose syntactic extensions for Prolog to expand its expressiveness, for instance to state logic formulas with their connectives verbatim. This allows to use all tools that were originally written for Prolog, for instance code linters and editors with syntax highlighting. We present the package library(plammar), a standard-compliant parser for Prolog source code, written in Prolog. It is able to automatically infer from example sentences the required operator definitions with their classes and precedences as well as the required Prolog language extensions. As a result, we can automatically answer the question: Is it possible to model these example sentences as valid Prolog clauses, and how?
We discuss and apply the two approaches to internal and external integrations for several domain-specific languages, namely the extended Backus–Naur form, GraphQL, XPath, and a controlled natural language to represent expert rules in if-then form. The created toolchain with library(dcg4pt) and library(plammar) yields new application opportunities for static Prolog source code analysis, which we also present.
In recent history, normalized digital surface models (nDSMs) have been constantly gaining importance as a means to solve large-scale geographic problems. High-resolution surface models are precious, as they can provide detailed information for a specific area. However, measurements with a high resolution are time consuming and costly. Only a few approaches exist to create high-resolution nDSMs for extensive areas. This article explores approaches to extract high-resolution nDSMs from low-resolution Sentinel-2 data, allowing us to derive large-scale models. We thereby utilize the advantages of Sentinel 2 being open access, having global coverage, and providing steady updates through a high repetition rate. Several deep learning models are trained to overcome the gap in producing high-resolution surface maps from low-resolution input data. With U-Net as a base architecture, we extend the capabilities of our model by integrating tailored multiscale encoders with differently sized kernels in the convolution as well as conformed self-attention inside the skip connection gates. Using pixelwise regression, our U-Net base models can achieve a mean height error of approximately 2 m. Moreover, through our enhancements to the model architecture, we reduce the model error by more than 7%.
An innovative framework has been developed for teamwork of two quadcopter formations, each having its specified formation geometry, assigned task, and matching control scheme. Position control for quadcopters in one of the formations has been implemented through a Linear Quadratic Regulator Proportional Integral (LQR PI) control scheme based on explicit model following scheme. Quadcopters in the other formation are controlled through LQR PI servomechanism control scheme. These two control schemes are compared in terms of their performance and control effort. Both formations are commanded by respective ground stations through virtual leaders. Quadcopters in formations are able to track desired trajectories as well as hovering at desired points for selected time duration. In case of communication loss between ground station and any of the quadcopters, the neighboring quadcopter provides the command data, received from the ground station, to the affected unit. Proposed control schemes have been validated through extensive simulations using MATLAB®/Simulink® that provided favorable results.
Lifetime techniques are applied to diverse fields of study including materials sciences, semiconductor physics, biology, molecular biophysics and photochemistry.
Here we present DDRS4PALS, a software for the acquisition and simulation of lifetime spectra using the DRS4 evaluation board (Paul Scherrer Institute, Switzerland) for time resolved measurements and digitization of detector output pulses. Artifact afflicted pulses can be corrected or rejected prior to the lifetime calculation to provide the generation of high-quality lifetime spectra, which are crucial for a profound analysis, i.e. the decomposition of the true information. Moreover, the pulses can be streamed on an (external) hard drive during the measurement and subsequently downloaded in the offline mode without being connected to the hardware. This allows the generation of various lifetime spectra at different configurations from one single measurement and, hence, a meaningful comparison in terms of analyzability and quality. Parallel processing and an integrated JavaScript based language provide convenient options to accelerate and automate time consuming processes such as lifetime spectra simulations.
This work proposes a novel approach to disperse dense transmission intervals and reduce bursty traffic patterns without the need for centralized control. Furthermore, by keeping the mechanism as close to the Long Range Wide Area Network (LoRaWAN) standard as possible the suggested mechanism can be deployed within existing networks and can even be co-deployed with other devices.
The DAEDALUS mission concept aims at exploring and characterising the entrance and initial part of Lunar lava tubes within a compact, tightly integrated spherical robotic device, with a complementary payload set and autonomous capabilities.
The mission concept addresses specifically the identification and characterisation of potential resources for future ESA exploration, the local environment of the subsurface and its geologic and compositional structure.
A sphere is ideally suited to protect sensors and scientific equipment in rough, uneven environments.
It will house laser scanners, cameras and ancillary payloads.
The sphere will be lowered into the skylight and will explore the entrance shaft, associated caverns and conduits. Lidar (light detection and ranging) systems produce 3D models with high spatial accuracy independent of lighting conditions and visible features.
Hence this will be the primary exploration toolset within the sphere.
The additional payload that can be accommodated in the robotic sphere consists of camera systems with panoramic lenses and scanners such as multi-wavelength or single-photon scanners.
A moving mass will trigger movements.
The tether for lowering the sphere will be used for data communication and powering the equipment during the descending phase.
Furthermore, the connector tether-sphere will host a WIFI access point, such that data of the conduit can be transferred to the surface relay station. During the exploration phase, the robot will be disconnected from the cable, and will use wireless communication.
Emergency autonomy software will ensure that in case of loss of communication, the robot will continue the nominal mission.
Mapping and localization of mobile robots in an unknown environment are essential for most high-level operations like autonomous navigation or exploration. This paper presents a novel approach for combining estimated trajectories, namely curvefusion. The robot used in the experiments is equipped with a horizontally mounted 2D profiler, a constantly spinning 3D laser scanner and a GPS module. The proposed algorithm first combines trajectories from different sensors to optimize poses of the planar three degrees of freedom (DoF) trajectory, which is then fed into continuous-time simultaneous localization and mapping (SLAM) to further improve the trajectory. While state-of-the-art multi-sensor fusion methods mainly focus on probabilistic methods, our approach instead adopts a deformation-based method to optimize poses. To this end, a similarity metric for curved shapes is introduced into the robotics community to fuse the estimated trajectories. Additionally, a shape-based point correspondence estimation method is applied to the multi-sensor time calibration. Experiments show that the proposed fusion method can achieve relatively better accuracy, even if the error of the trajectory before fusion is large, which demonstrates that our method can still maintain a certain degree of accuracy in an environment where typical pose estimation methods have poor performance. In addition, the proposed time-calibration method also achieves high accuracy in estimating point correspondences.
In recent years several community testbeds as well as participatory sensing platforms have successfully established themselves to provide open data to everyone interested. Each of them with a specific goal in mind, ranging from collecting radio coverage data up to environmental and radiation data. Such data can be used by the community in their decision making, whether to subscribe to a specific mobile phone service that provides good coverage in an area or in finding a sunny and warm region for the summer holidays.
However, the existing platforms are usually limiting themselves to directly measurable network QoS. If such a crowdsourced data set provides more in-depth derived measures, this would enable an even better decision making. A community-driven crowdsensing platform that derives spatial application-layer user experience from resource-friendly bandwidth estimates would be such a case, video streaming services come to mind as a prime example. In this paper we present a concept for such a system based on an initial prototype that eases the collection of data necessary to determine mobile-specific QoE at large scale. In addition we reason why the simple quality metric proposed here can hold its own.
In many cases, problems, data, or information can be modeled as graphs. Graphs can be used as a tool for modeling in any case where connections between distinguishable objects occur. Any graph consists of a set of objects, called vertices, and a set of connections, called edges, such that any edge connects a pair of vertices. For example, a social network can be modeled by a graph by
transforming the users of the network into vertices and friendship relations between users into edges. Also physical networks like computer networks or transportation networks, for example, the metro network of a city, can be seen as graphs.
For making graphs and, thereby, the data that is modeled, well-understandable for users, we need a visualization. Graph drawing deals with algorithms for visualizing graphs. In this thesis, especially the use of crossings and curves is investigated for graph drawing problems under additional constraints. The constraints that occur in the problems investigated in this thesis especially restrict the positions of (a part of) the vertices; this is done either as a hard constraint or as an optimization criterion.
Cover contact graphs
(2012)
We study problems that arise in the context of covering certain geometric objects called seeds (e.g., points or disks) by a set of other geometric objects called cover (e.g., a set of disks or homothetic triangles). We insist that the interiors of the seeds and the cover elements are pairwise disjoint, respectively, but they can touch. We call the contact graph of a cover a cover contact graph (CCG). We are interested in three types of tasks, both in the general case and in the special case of seeds on a line: (a) deciding whether a given seed set has a connected CCG, (b) deciding whether a given graph has a realization as a CCG on a given seed set, and (c) bounding the sizes of certain classes of CCG’s. Concerning (a) we give efficient algorithms for the case that seeds are points and show that the problem becomes hard if seeds and covers are disks. Concerning (b) we show that this problem is hard even for point seeds and disk covers (given a fixed correspondence between graph vertices and seeds). Concerning (c) we obtain upper and lower bounds on the number of CCG’s for point seeds.
Climate models are the tool of choice for scientists researching climate change. Like all models they suffer from errors, particularly systematic and location-specific representation errors. One way to reduce these errors is model output statistics (MOS) where the model output is fitted to observational data with machine learning. In this work, we assess the use of convolutional Deep Learning climate MOS approaches and present the ConvMOS architecture which is specifically designed based on the observation that there are systematic and location-specific errors in the precipitation estimates of climate models. We apply ConvMOS models to the simulated precipitation of the regional climate model REMO, showing that a combination of per-location model parameters for reducing location-specific errors and global model parameters for reducing systematic errors is indeed beneficial for MOS performance. We find that ConvMOS models can reduce errors considerably and perform significantly better than three commonly used MOS approaches and plain ResNet and U-Net models in most cases. Our results show that non-linear MOS models underestimate the number of extreme precipitation events, which we alleviate by training models specialized towards extreme precipitation events with the imbalanced regression method DenseLoss. While we consider climate MOS, we argue that aspects of ConvMOS may also be beneficial in other domains with geospatial data, such as air pollution modeling or weather forecasts.
This article presents a novel method for controlling a virtual audience system (VAS) in Virtual Reality (VR) application, called STAGE, which has been originally designed for supervised public speaking training in university seminars dedicated to the preparation and delivery of scientific talks. We are interested in creating pedagogical narratives: narratives encompass affective phenomenon and rather than organizing events changing the course of a training scenario, pedagogical plans using our system focus on organizing the affects it arouses for the trainees. Efficiently controlling a virtual audience towards a specific training objective while evaluating the speaker’s performance presents a challenge for a seminar instructor: the high level of cognitive and physical demands required to be able to control the virtual audience, whilst evaluating speaker’s performance, adjusting and allowing it to quickly react to the user’s behaviors and interactions. It is indeed a critical limitation of a number of existing systems that they rely on a Wizard of Oz approach, where the tutor drives the audience in reaction to the user’s performance. We address this problem by integrating with a VAS a high-level control component for tutors, which allows using predefined audience behavior rules, defining custom ones, as well as intervening during run-time for finer control of the unfolding of the pedagogical plan. At its core, this component offers a tool to program, select, modify and monitor interactive training narratives using a high-level representation. The STAGE offers the following features: i) a high-level API to program pedagogical narratives focusing on a specific public speaking situation and training objectives, ii) an interactive visualization interface iii) computation and visualization of user metrics, iv) a semi-autonomous virtual audience composed of virtual spectators with automatic reactions to the speaker and surrounding spectators while following the pedagogical plan V) and the possibility for the instructor to embody a virtual spectator to ask questions or guide the speaker from within the Virtual Environment. We present here the design, and implementation of the tutoring system and its integration in STAGE, and discuss its reception by end-users.
Constraining graph layouts - that is, restricting the placement of vertices and the routing of edges to obey certain constraints - is common practice in graph drawing.
In this book, we discuss algorithmic results on two different restriction types:
placing vertices on the outer face and on the integer grid.
For the first type, we look into the outer k-planar and outer k-quasi-planar graphs, as well as giving a linear-time algorithm to recognize full and closed outer k-planar graphs Monadic Second-order Logic.
For the second type, we consider the problem of transferring a given planar drawing onto the integer grid while perserving the original drawings topology;
we also generalize a variant of Cauchy's rigidity theorem for orthogonal polyhedra of genus 0 to those of arbitrary genus.
Presence is often considered the most important quale describing the subjective feeling of being in a computer-generated and/or computer-mediated virtual environment. The identification and separation of orthogonal presence components, i.e., the place illusion and the plausibility illusion, has been an accepted theoretical model describing Virtual Reality (VR) experiences for some time. This perspective article challenges this presence-oriented VR theory. First, we argue that a place illusion cannot be the major construct to describe the much wider scope of virtual, augmented, and mixed reality (VR, AR, MR: or XR for short). Second, we argue that there is no plausibility illusion but merely plausibility, and we derive the place illusion caused by the congruent and plausible generation of spatial cues and similarly for all the current model’s so-defined illusions. Finally, we propose congruence and plausibility to become the central essential conditions in a novel theoretical model describing XR experiences and effects.
Computing Generic Causes of Revelation of the Quranic Verses Using Machine Learning Techniques
(2011)
Because many verses of the holy Quran are similar, there is high probability that, similar verses addressing same issues share same generic causes of revelation. In this study, machine learning techniques have been employed in order to automatically derive causes of revelation of Quranic verses. The derivation of the causes of revelation is viewed as a classification problem. Initially the categories are based on the verses with known causes of revelation, and the testing set consists of the remaining verses. Based on a computed threshold value, a naïve Bayesian classifier is used to categorize some verses. After that, using a decision tree classifier the remaining uncategorized verses are separated into verses that contain indicators (resultative connectors, causative expressions…), and those that do not. As for those verses having indicators, each one is segmented into its constituent clauses by identification of the linking indicators. Then a dominant clause is extracted and considered either as the cause of revelation, or post-processed by adding or subtracting some terms to form a causal clause that constitutes the cause of revelation. Concerning remaining unclassified verses without indicators, a naive Bayesian classifier is again used to assign each one of them to one of the existing classes based on features and topics similarity. As for verses that could not be classified so far, manual classification was made by considering each verse as a category on its own. The result obtained in this study is encouraging, and shows that automatic derivation of Quranic verses’ generic causes of revelation is achievable, and reasonably reliable for understanding and implementing the teachings of the Quran.
Given a collection of diverging documents about some lost original text, any person interested in the text would try reconstructing it from the diverging documents. Whether it is eclecticism, stemmatics, or copy-text, one is expected to explicitly or indirectly select one of the documents as a starting point or as a base text, which could be emended through comparison with remaining documents, so that a text that could be designated as the original document is generated. Unfortunately the process of giving priority to one of the documents also known as witnesses is a subjective approach. In fact even Cladistics, which could be considered as a computer-based approach of implementing stemmatics, does not present or recommend users to select a certain witness as a starting point for the process of reconstructing the original document. In this study, a computational method using a rule-based Bayesian classifier is used, to assist text scholars in their attempts of reconstructing a non-existing document from some available witnesses. The method developed in this study consists of selecting a base text successively and collating it with remaining documents. Each completed collation cycle stores the selected base text and its closest witness, along with a weighted score of their similarities and differences. At the end of the collation process, a witness selected more often by majority of base texts is considered as the probable base text of the collection. Witnesses’ scores are weighted using a weighting system, based on effects of types of textual modifications on the process of reconstructing original documents. Users have the possibility to select between baseless and base text collation. If a base text is selected, the task is reduced to ranking the witnesses with respect to the base text, otherwise a base text as well as ranking of the witnesses with respect to the base text are computed and displayed on a bar diagram. Additionally this study includes a recursive algorithm for automatically reconstructing the original text from the identified base text and ranked witnesses.
To enable a sustainable supply of chemicals, novel biotechnological solutions are required that replace the reliance on fossil resources. One potential solution is to utilize tailored biosynthetic modules for the metabolic conversion of CO2 or organic waste to chemicals and fuel by microorganisms. Currently, it is challenging to commercialize biotechnological processes for renewable chemical biomanufacturing because of a lack of highly active and specific biocatalysts. As experimental methods to engineer biocatalysts are time- and cost-intensive, it is important to establish efficient and reliable computational tools that can speed up the identification or optimization of selective, highly active, and stable enzyme variants for utilization in the biotechnological industry. Here, we review and suggest combinations of effective state-of-the-art software and online tools available for computational enzyme engineering pipelines to optimize metabolic pathways for the biosynthesis of renewable chemicals. Using examples relevant for biotechnology, we explain the underlying principles of enzyme engineering and design and illuminate future directions for automated optimization of biocatalysts for the assembly of synthetic metabolic pathways.
Complexity and Partitions
(2001)
Computational complexity theory usually investigates the complexity of sets, i.e., the complexity of partitions into two parts. But often it is more appropriate to represent natural problems by partitions into more than two parts. A particularly interesting class of such problems consists of classification problems for relations. For instance, a binary relation R typically defines a partitioning of the set of all pairs (x,y) into four parts, classifiable according to the cases where R(x,y) and R(y,x) hold, only R(x,y) or only R(y,x) holds or even neither R(x,y) nor R(y,x) is true. By means of concrete classification problems such as Graph Embedding or Entailment (for propositional logic), this thesis systematically develops tools, in shape of the boolean hierarchy of NP-partitions and its refinements, for the qualitative analysis of the complexity of partitions generated by NP-relations. The Boolean hierarchy of NP-partitions is introduced as a generalization of the well-known and well-studied Boolean hierarchy (of sets) over NP. Whereas the latter hierarchy has a very simple structure, the situation is much more complicated for the case of partitions into at least three parts. To get an idea of this hierarchy, alternative descriptions of the partition classes are given in terms of finite, labeled lattices. Based on these characterizations the Embedding Conjecture is established providing the complete information on the structure of the hierarchy. This conjecture is supported by several results. A natural extension of the Boolean hierarchy of NP-partitions emerges from the lattice-characterization of its classes by considering partition classes generated by finite, labeled posets. It turns out that all significant ideas translate from the case of lattices. The induced refined Boolean hierarchy of NP-partitions enables us more accuratly capturing the complexity of certain relations (such as Graph Embedding) and a description of projectively closed partition classes.
We consider competitive location problems where two competing providers place their facilities sequentially and users can decide between the competitors. We assume that both competitors act non-cooperatively and aim at maximizing their own benefits. We investigate the complexity and approximability of such problems on graphs, in particular on simple graph classes such as trees and paths. We also develop fast algorithms for single competitive location problems where each provider places a single facilty. Voting location, in contrast, aims at identifying locations that meet social criteria. The provider wants to satisfy the users (customers) of the facility to be opened. In general, there is no location that is favored by all users. Therefore, a satisfactory compromise has to be found. To this end, criteria arising from voting theory are considered. The solution of the location problem is understood as the winner of a virtual election among the users of the facilities, in which the potential locations play the role of the candidates and the users represent the voters. Competitive and voting location problems turn out to be closely related.
Scalability is often mentioned in literature, but a stringent definition is missing. In particular, there is no general scalability assessment which clearly indicates whether a system scales or not or whether a system scales better than another. The key contribution of this article is the definition of a scalability index (SI) which quantifies if a system scales in comparison to another system, a hypothetical system, e.g., linear system, or the theoretically optimal system. The suggested SI generalizes different metrics from literature, which are specialized cases of our SI. The primary target of our scalability framework is, however, benchmarking of two systems, which does not require any reference system. The SI is demonstrated and evaluated for different use cases, that are (1) the performance of an IoT load balancer depending on the system load, (2) the availability of a communication system depending on the size and structure of the network, (3) scalability comparison of different location selection mechanisms in fog computing with respect to delays and energy consumption; (4) comparison of time-sensitive networking (TSN) mechanisms in terms of efficiency and utilization. Finally, we discuss how to use and how not to use the SI and give recommendations and guidelines in practice. To the best of our knowledge, this is the first work which provides a general SI for the comparison and benchmarking of systems, which is the primary target of our scalability analysis.
Tardigrades have fascinated researchers for more than 300 years because of their extraordinary capability to undergo cryptobiosis and survive extreme environmental conditions. However, the survival mechanisms of tardigrades are still poorly understood mainly due to the absence of detailed knowledge about the proteome and genome of these organisms. Our study was intended to provide a basis for the functional characterization of expressed proteins in different states of tardigrades. High-throughput, high-accuracy proteomics in combination with a newly developed tardigrade specific protein database resulted in the identification of more than 3000 proteins in three different states: early embryonic state and adult animals in active and anhydrobiotic state. This comprehensive proteome resource includes protein families such as chaperones, antioxidants, ribosomal proteins, cytoskeletal proteins, transporters, protein channels, nutrient reservoirs, and developmental proteins. A comparative analysis of protein families in the different states was performed by calculating the exponentially modified protein abundance index which classifies proteins in major and minor components. This is the first step to analyzing the proteins involved in early embryonic development, and furthermore proteins which might play an important role in the transition into the anhydrobiotic state.
Today’s advanced Internet-of-Things applications raise technical challenges on cloud, edge, and fog computing. The design of an efficient, virtualized, context-aware, self-configuring orchestration system of a fog computing system constitutes a major development effort within this very innovative area of research. In this paper we describe the architecture and relevant implementation aspects of a cloudless resource monitoring system interworking with an SDN/NFV infrastructure. It realizes the basic monitoring component of the fundamental MAPE-K principles employed in autonomic computing. Here we present the hierarchical layering and functionality within the underlying fog nodes to generate a working prototype of an intelligent, self-managed orchestrator for advanced IoT applications and services. The latter system has the capability to monitor automatically various performance aspects of the resource allocation among multiple hosts of a fog computing system interconnected by SDN.
CLIP knows image aesthetics
(2022)
Most Image Aesthetic Assessment (IAA) methods use a pretrained ImageNet classification model as a base to fine-tune. We hypothesize that content classification is not an optimal pretraining task for IAA, since the task discourages the extraction of features that are useful for IAA, e.g., composition, lighting, or style. On the other hand, we argue that the Contrastive Language-Image Pretraining (CLIP) model is a better base for IAA models, since it has been trained using natural language supervision. Due to the rich nature of language, CLIP needs to learn a broad range of image features that correlate with sentences describing the image content, composition, environments, and even subjective feelings about the image. While it has been shown that CLIP extracts features useful for content classification tasks, its suitability for tasks that require the extraction of style-based features like IAA has not yet been shown. We test our hypothesis by conducting a three-step study, investigating the usefulness of features extracted by CLIP compared to features obtained from the last layer of a comparable ImageNet classification model. In each step, we get more computationally expensive. First, we engineer natural language prompts that let CLIP assess an image's aesthetic without adjusting any weights in the model. To overcome the challenge that CLIP's prompting only is applicable to classification tasks, we propose a simple but effective strategy to convert multiple prompts to a continuous scalar as required when predicting an image's mean aesthetic score. Second, we train a linear regression on the AVA dataset using image features obtained by CLIP's image encoder. The resulting model outperforms a linear regression trained on features from an ImageNet classification model. It also shows competitive performance with fully fine-tuned networks based on ImageNet, while only training a single layer. Finally, by fine-tuning CLIP's image encoder on the AVA dataset, we show that CLIP only needs a fraction of training epochs to converge, while also performing better than a fine-tuned ImageNet model. Overall, our experiments suggest that CLIP is better suited as a base model for IAA methods than ImageNet pretrained networks.
The charged aerosol detector (CAD) is the latest representative of aerosol-based detectors that generate a response independent of the analytes' chemical structure. This study was aimed at accurately predicting the CAD response of homologous fatty acids under varying experimental conditions. Fatty acids from C12 to C18 were used as model substances due to semivolatile characterics that caused non-uniform CAD behaviour. Considering both experimental conditions and molecular descriptors, a mixed quantitative structure-property relationship (QSPR) modeling was performed using Gradient Boosted Trees (GBT). The ensemble of 10 decisions trees (learning rate set at 0.55, the maximal depth set at 5, and the sample rate set at 1.0) was able to explain approximately 99% (Q\(^2\): 0.987, RMSE: 0.051) of the observed variance in CAD responses. Validation using an external test compound confirmed the high predictive ability of the model established (R-2: 0.990, RMSEP: 0.050). With respect to the intrinsic attribute selection strategy, GBT used almost all independent variables during model building. Finally, it attributed the highest importance to the power function value, the flow rate of the mobile phase, evaporation temperature, the content of the organic solvent in the mobile phase and the molecular descriptors such as molecular weight (MW), Radial Distribution Function-080/weighted by mass (RDF080m) and average coefficient of the last eigenvector from distance/detour matrix (Ve2_D/Dt). The identification of the factors most relevant to the CAD responsiveness has contributed to a better understanding of the underlying mechanisms of signal generation. An increased CAD response that was obtained for acetone as organic modifier demonstrated its potential to replace the more expensive and environmentally harmful acetonitrile.
The ecosystem of the high northern latitudes is affected by the recently changing environmental conditions. The Arctic has undergone a significant climatic change over the last decades. The land coverage is changing and a phenological response to the warming is apparent. Remotely sensed data can assist the monitoring and quantification of these changes. The remote sensing of the Arctic was predominantly carried out by the usage of optical sensors but these encounter problems in the Arctic environment, e.g. the frequent cloud cover or the solar geometry. In contrast, the imaging of Synthetic Aperture Radar is not affected by the cloud cover and the acquisition of radar imagery is independent of the solar illumination. The objective of this work was to explore how polarimetric Synthetic Aperture Radar (PolSAR) data of TerraSAR-X, TanDEM-X, Radarsat-2 and ALOS PALSAR and interferometric-derived digital elevation model data of the TanDEM-X Mission can contribute to collect meaningful information on the actual state of the Arctic Environment. The study was conducted for Canadian sites of the Mackenzie Delta Region and Banks Island and in situ reference data were available for the assessment. The up-to-date analysis of the PolSAR data made the application of the Non-Local Means filtering and of the decomposition of co-polarized data necessary.
The Non-Local Means filter showed a high capability to preserve the image values, to keep the edges and to reduce the speckle. This supported not only the suitability for the interpretation but also for the classification. The classification accuracies of Non-Local Means filtered data were in average +10% higher compared to unfiltered images. The correlation of the co- and quad-polarized decomposition features was high for classes with distinct surface or double bounce scattering and a usage of the co-polarized data is beneficial for regions of natural land coverage and for low vegetation formations with little volume scattering. The evaluation further revealed that the X- and C-Band were most sensitive to the generalized land cover classes. It was found that the X-Band data were sensitive to low vegetation formations with low shrub density, the C-Band data were sensitive to the shrub density and the shrub dominated tundra. In contrast, the L-Band data were less sensitive to the land cover. Among the different dual-polarized data the HH/VV-polarized data were identified to be most meaningful for the characterization and classification, followed by the HH/HV-polarized and the VV/VH-polarized data. The quad-polarized data showed highest sensitivity to the land cover but differences to the co-polarized data were small. The accuracy assessment showed that spectral information was required for accurate land cover classification. The best results were obtained when spectral and radar information was combined. The benefit of including radar data in the classification was up to +15% accuracy and most significant for the classes wetland and sparse vegetated tundra. The best classifications were realized with quad-polarized C-Band and multispectral data and with co-polarized X-Band and multispectral data. The overall accuracy was up to 80% for unsupervised and up to 90% for supervised classifications. The results indicated that the shortwave co-polarized data show promise for the classification of tundra land cover since the polarimetric information is sensitive to low vegetation and the wetlands. Furthermore, co-polarized data provide a higher spatial resolution than the quad-polarized data.
The analysis of the intermediate digital elevation model data of the TanDEM-X showed a high potential for the characterization of the surface morphology. The basic and relative topographic features were shown to be of high relevance for the quantification of the surface morphology and an area-wide application is feasible. In addition, these data were of value for the classification and delineation of landforms. Such classifications will assist the delineation of geomorphological units and have potential to identify locations of actual and future morphologic activity.
The emerging serverless computing may meet Edge Cloud in a beneficial manner as the two offer flexibility and dynamicity in optimizing finite hardware resources. However, the lack of proper study of a joint platform leaves a gap in literature about consumption and performance of such integration. To this end, this paper identifies the key questions and proposes a methodology to answer them.
Mindfulness is considered an important factor of an individual's subjective well-being. Consequently, Human-Computer Interaction (HCI) has investigated approaches that strengthen mindfulness, i.e., by inventing multimedia technologies to support mindfulness meditation. These approaches often use smartphones, tablets, or consumer-grade desktop systems to allow everyday usage in users' private lives or in the scope of organized therapies. Virtual, Augmented, and Mixed Reality (VR, AR, MR; in short: XR) significantly extend the design space for such approaches. XR covers a wide range of potential sensory stimulation, perceptive and cognitive manipulations, content presentation, interaction, and agency. These facilities are linked to typical XR-specific perceptions that are conceptually closely related to mindfulness research, such as (virtual) presence and (virtual) embodiment. However, a successful exploitation of XR that strengthens mindfulness requires a systematic analysis of the potential interrelation and influencing mechanisms between XR technology, its properties, factors, and phenomena and existing models and theories of the construct of mindfulness. This article reports such a systematic analysis of XR-related research from HCI and life sciences to determine the extent to which existing research frameworks on HCI and mindfulness can be applied to XR technologies, the potential of XR technologies to support mindfulness, and open research gaps. Fifty papers of ACM Digital Library and National Institutes of Health's National Library of Medicine (PubMed) with and without empirical efficacy evaluation were included in our analysis. The results reveal that at the current time, empirical research on XR-based mindfulness support mainly focuses on therapy and therapeutic outcomes. Furthermore, most of the currently investigated XR-supported mindfulness interactions are limited to vocally guided meditations within nature-inspired virtual environments. While an analysis of empirical research on those systems did not reveal differences in mindfulness compared to non-mediated mindfulness practices, various design proposals illustrate that XR has the potential to provide interactive and body-based innovations for mindfulness practice. We propose a structured approach for future work to specify and further explore the potential of XR as mindfulness-support. The resulting framework provides design guidelines for XR-based mindfulness support based on the elements and psychological mechanisms of XR interactions.
This article presents an immersive virtual reality (VR) system for training classroom management skills, with a specific focus on learning to manage disruptive student behavior in face-to-face, one-to-many teaching scenarios. The core of the system is a real-time 3D virtual simulation of a classroom populated by twenty-four semi-autonomous virtual students. The system has been designed as a companion tool for classroom management seminars in a syllabus for primary and secondary school teachers. This will allow lecturers to link theory with practice using the medium of VR. The system is therefore designed for two users: a trainee teacher and an instructor supervising the training session. The teacher is immersed in a real-time 3D simulation of a classroom by means of a head-mounted display and headphone. The instructor operates a graphical desktop console, which renders a view of the class and the teacher whose avatar movements are captured by a marker less tracking system. This console includes a 2D graphics menu with convenient behavior and feedback control mechanisms to provide human-guided training sessions. The system is built using low-cost consumer hardware and software. Its architecture and technical design are described in detail. A first evaluation confirms its conformance to critical usability requirements (i.e., safety and comfort, believability, simplicity, acceptability, extensibility, affordability, and mobility). Our initial results are promising and constitute the necessary first step toward a possible investigation of the efficiency and effectiveness of such a system in terms of learning outcomes and experience.
The Internet of Things (IoT) enables a variety of smart applications, including smart home, smart manufacturing, and smart city. By enhancing Business Process Management Systems with IoT capabilities, the execution and monitoring of business processes can be significantly improved. Providing a holistic support for modeling, executing and monitoring IoT-driven processes, however, constitutes a challenge. Existing process modeling and process execution languages, such as BPMN 2.0, are unable to fully meet the IoT characteristics (e.g., asynchronicity and parallelism) of IoT-driven processes. In this article, we present BPMNE4IoT—A holistic framework for modeling, executing and monitoring IoT-driven processes. We introduce various artifacts and events based on the BPMN 2.0 metamodel that allow realizing the desired IoT awareness of business processes. The framework is evaluated along two real-world scenarios from two different domains. Moreover, we present a user study for comparing BPMNE4IoT and BPMN 2.0. In particular, this study has confirmed that the BPMNE4IoT framework facilitates the support of IoT-driven processes.
This paper describes the estimation of the body weight of a person in front of an RGB-D camera. A survey of different methods for body weight estimation based on depth sensors is given. First, an estimation of people standing in front of a camera is presented. Second, an approach based on a stream of depth images is used to obtain the body weight of a person walking towards a sensor. The algorithm first extracts features from a point cloud and forwards them to an artificial neural network (ANN) to obtain an estimation of body weight. Besides the algorithm for the estimation, this paper further presents an open-access dataset based on measurements from a trauma room in a hospital as well as data from visitors of a public event. In total, the dataset contains 439 measurements. The article illustrates the efficiency of the approach with experiments with persons lying down in a hospital, standing persons, and walking persons. Applicable scenarios for the presented algorithm are body weight-related dosing of emergency patients.
Cosmology often uses intricate formulas and mathematics to derive new theories and concepts. We do something different in this paper: We look at biological processes and derive from these heuristics so that the revised cosmology agrees with astronomical observations but does also agree with standard biological observations. We show that we then have to replace any type of singularity at the start of the universe by a condensation nucleus and that the very early period of the universe usually assumed to be inflation has to be replaced by a period of rapid crystal growth as in Weiss magnetization domains.
Impressively, these minor modifications agree well with astronomical observations including removing the strong inflation perturbations which were never observed in the recent BICEP2 experiments. Furthermore, looking at biological principles suggests that such a new theory with a condensation nucleus at start and a first rapid phase of magnetization-like growth of the ordered, physical laws obeying lattice we live in is in fact the only convincing theory of the early phases of our universe that also is compatible with current observations.
We show in detail in the following that such a process of crystal creation, breaking of new crystal seeds and ultimate evaporation of the present crystal readily leads over several generations to an evolution and selection of better, more stable and more self-organizing crystals. Moreover, this explains the “fine-tuning” question why our universe is fine-tuned to favor life: Our Universe is so self-organizing to have enough offspring and the detailed physics involved is at the same time highly favorable for all self-organizing processes including life.
This biological theory contrasts with current standard inflation cosmologies. The latter do not perform well in explaining any phenomena of sophisticated structure creation or self-organization. As proteins can only thermodynamically fold by increasing the entropy in the solution around them we suggest for cosmology a condensation nucleus for a universe can form only in a “chaotic ocean” of string-soup or quantum foam if the entropy outside of the nucleus rapidly increases. We derive an interaction potential for 1 to n-dimensional strings or quantum-foams and show that they allow only 1D, 2D, 4D or octonion interactions. The latter is the richest structure and agrees to the E8 symmetry fundamental to particle physics and also compatible with the ten dimensional string theory E8 which is part of the M-theory. Interestingly, any other interactions of other dimensionality can be ruled out using Hurwitz compositional theorem. Crystallization explains also extremely well why we have only one macroscopic reality and where the worldlines of alternative trajectories exist: They are in other planes of the crystal and for energy reasons they crystallize mostly at the same time, yielding a beautiful and stable crystal. This explains decoherence and allows to determine the size of Planck´s quantum h (very small as separation of crystal layers by energy is extremely strong).
Ultimate dissolution of real crystals suggests an explanation for dark energy agreeing with estimates for the “big rip”. The halo distribution of dark matter favoring galaxy formation is readily explained by a crystal seed starting with unit cells made of normal and dark matter.
That we have only matter and not antimatter can be explained as there may be right handed mattercrystals and left-handed antimatter crystals. Similarly, real crystals are never perfect and we argue that exactly such irregularities allow formation of galaxies, clusters and superclusters. Finally, heuristics from genetics suggest to look for a systems perspective to derive correct vacuum and Higgs Boson energies.
The design and evaluation of assisting technologies to support behavior change processes have become an essential topic within the field of human-computer interaction research in general and the field of immersive intervention technologies in particular. The mechanisms and success of behavior change techniques and interventions are broadly investigated in the field of psychology. However, it is not always easy to adapt these psychological findings to the context of immersive technologies. The lack of theoretical foundation also leads to a lack of explanation as to why and how immersive interventions support behavior change processes. The Behavioral Framework for immersive Technologies (BehaveFIT) addresses this lack by 1) presenting an intelligible categorization and condensation of psychological barriers and immersive features, by 2) suggesting a mapping that shows why and how immersive technologies can help to overcome barriers and finally by 3) proposing a generic prediction path that enables a structured, theory-based approach to the development and evaluation of immersive interventions. These three steps explain how BehaveFIT can be used, and include guiding questions for each step. Further, two use cases illustrate the usage of BehaveFIT. Thus, the present paper contributes to guidance for immersive intervention design and evaluation, showing that immersive interventions support behavior change processes and explain and predict 'why' and 'how' immersive interventions can bridge the intention-behavior-gap.
The development of ICT infrastructures has facilitated the emergence of new paradigms for looking at society and the environment over the last few years. Participatory environmental sensing, i.e. directly involving citizens in environmental monitoring, is one example, which is hoped to encourage learning and enhance awareness of environmental issues. In this paper, an analysis of the behaviour of individuals involved in noise sensing is presented. Citizens have been involved in noise measuring activities through the WideNoise smartphone application. This application has been designed to record both objective (noise samples) and subjective (opinions, feelings) data. The application has been open to be used freely by anyone and has been widely employed worldwide. In addition, several test cases have been organised in European countries. Based on the information submitted by users, an analysis of emerging awareness and learning is performed. The data show that changes in the way the environment is perceived after repeated usage of the application do appear. Specifically, users learn how to recognise different noise levels they are exposed to. Additionally, the subjective data collected indicate an increased user involvement in time and a categorisation effect between pleasant and less pleasant environments.
Background
Colorectal cancer is a leading cause of cancer-related deaths worldwide. The best method to prevent CRC is a colonoscopy. However, not all colon polyps have the risk of becoming cancerous. Therefore, polyps are classified using different classification systems. After the classification, further treatment and procedures are based on the classification of the polyp. Nevertheless, classification is not easy. Therefore, we suggest two novel automated classifications system assisting gastroenterologists in classifying polyps based on the NICE and Paris classification.
Methods
We build two classification systems. One is classifying polyps based on their shape (Paris). The other classifies polyps based on their texture and surface patterns (NICE). A two-step process for the Paris classification is introduced: First, detecting and cropping the polyp on the image, and secondly, classifying the polyp based on the cropped area with a transformer network. For the NICE classification, we design a few-shot learning algorithm based on the Deep Metric Learning approach. The algorithm creates an embedding space for polyps, which allows classification from a few examples to account for the data scarcity of NICE annotated images in our database.
Results
For the Paris classification, we achieve an accuracy of 89.35 %, surpassing all papers in the literature and establishing a new state-of-the-art and baseline accuracy for other publications on a public data set. For the NICE classification, we achieve a competitive accuracy of 81.13 % and demonstrate thereby the viability of the few-shot learning paradigm in polyp classification in data-scarce environments. Additionally, we show different ablations of the algorithms. Finally, we further elaborate on the explainability of the system by showing heat maps of the neural network explaining neural activations.
Conclusion
Overall we introduce two polyp classification systems to assist gastroenterologists. We achieve state-of-the-art performance in the Paris classification and demonstrate the viability of the few-shot learning paradigm in the NICE classification, addressing the prevalent data scarcity issues faced in medical machine learning.
In this paper we study connectivity augmentation problems. Given a connected graph G with some desirable property, we want to make G 2-vertex connected (or 2-edge connected) by adding edges such that the resulting graph keeps the property. The aim is to add as few edges as possible. The property that we consider is planarity, both in an abstract graph-theoretic and in a geometric setting, where vertices correspond to points in the plane and edges to straight-line segments.
We show that it is NP-hard to nd a minimum-cardinality augmentation that makes a planar graph 2-edge connected. For making a planar graph 2-vertex connected this was known. We further show that both problems are hard in the geometric setting, even when restricted to trees. The problems remain hard for higher degrees of connectivity. On the other hand we give polynomial-time algorithms for the special case of convex geometric graphs.
We also study the following related problem. Given a planar (plane geometric) graph G, two vertices s and t of G, and an integer c, how many edges have to be added to G such that G is still planar (plane geometric) and contains c edge- (or vertex-) disjoint s{t paths? For the planar case we give a linear-time algorithm for c = 2. For the plane geometric case we give optimal worst-case bounds for c = 2; for c = 3 we characterize the cases that have a solution.