004 Datenverarbeitung; Informatik
Refine
Has Fulltext
- yes (257)
Year of publication
Document Type
- Journal article (123)
- Doctoral Thesis (68)
- Working Paper (37)
- Preprint (19)
- Conference Proceeding (8)
- Report (2)
Language
- English (257) (remove)
Keywords
- virtual reality (16)
- Datennetz (14)
- Leistungsbewertung (13)
- Quran (8)
- Robotik (8)
- Koran (7)
- Text Mining (7)
- Mobiler Roboter (6)
- Autonomer Roboter (5)
- Komplexitätstheorie (5)
- Maschinelles Lernen (5)
- Netzwerk (5)
- Optimierung (5)
- P4 (5)
- Theoretische Informatik (5)
- Visualisierung (5)
- artificial intelligence (5)
- machine learning (5)
- Overlay-Netz (4)
- Routing (4)
- SDN (4)
- Verteiltes System (4)
- XML (4)
- augmented reality (4)
- deep learning (4)
- human-computer interaction (4)
- Algorithmus (3)
- Approximationsalgorithmus (3)
- Bayesian classifier (3)
- Computer Vision (3)
- Deep learning (3)
- Dienstgüte (3)
- Drahtloses Sensorsystem (3)
- Graph (3)
- Graphenzeichnen (3)
- IoT (3)
- Komplexität (3)
- Machine Learning (3)
- Mensch-Maschine-Schnittstelle (3)
- Netzwerkmanagement (3)
- Optimization (3)
- Peer-to-Peer-Netz (3)
- Performance Evaluation (3)
- QoE (3)
- Quadrocopter (3)
- Quality of Experience (3)
- Rechnernetz (3)
- Ressourcenmanagement (3)
- Robotics (3)
- Softwarearchitektur (3)
- Textvergleich (3)
- Visualization (3)
- Wissensmanagement (3)
- approximation algorithm (3)
- crowdsensing (3)
- database (3)
- graph drawing (3)
- immersion (3)
- mHealth (3)
- neural networks (3)
- resistance (3)
- simulation (3)
- 5G (2)
- Algorithmische Geometrie (2)
- Ausfallsicheres System (2)
- Ausfallsicherheit (2)
- Base text (2)
- Benutzerschnittstelle (2)
- CSS (2)
- Cascading Style Sheets (2)
- Content Management (2)
- Crowdsourcing (2)
- Data Mining (2)
- Dot-Depth Problem (2)
- Drahtloses lokales Netz (2)
- Effizienter Algorithmus (2)
- Entscheidbarkeit (2)
- Fernwartung (2)
- Future Internet (2)
- Gothenburg model (2)
- Human-Robot-Interaction (2)
- IEEE 802.11 (2)
- Internet of Things (2)
- Kleinsatellit (2)
- Knowledge Management (2)
- Kreuzung (2)
- Localization (2)
- Lokalisation (2)
- MP-DCCP (2)
- Maschinelles Sehen (2)
- Mensch-Maschine-System (2)
- Mensch-Roboter-Interaktion (2)
- Meta-model (2)
- Mixed Reality (2)
- Modellierung (2)
- NP-hardness (2)
- PROLOG <Programmiersprache> (2)
- Punktwolke (2)
- Quadrotor (2)
- Resilience (2)
- Resource Management (2)
- Simulation (2)
- Situation Awareness (2)
- Software Defined Networking (2)
- Software Engineering (2)
- Teleoperation (2)
- Text mining (2)
- Textual alterations weighting system (2)
- Textual document collation (2)
- Theoretical Computer Science (2)
- User Interface (2)
- Verbotsmuster (2)
- Wissensrepräsentation (2)
- Wrapper <Programmierung> (2)
- XR (2)
- automation (2)
- connected mobility applications (2)
- cosmology (2)
- crossing minimization (2)
- decidability (2)
- design (2)
- dot-depth problem (2)
- education (2)
- educational tool (2)
- endliche Automaten (2)
- endoscopy (2)
- engineering (2)
- evolution (2)
- exposure (2)
- finite automata (2)
- fog computing (2)
- forbidden patterns (2)
- framework (2)
- fully convolutional neural networks (2)
- games (2)
- gastroenterology (2)
- genetics (2)
- historical document analysis (2)
- immersive technologies (2)
- knowledge representation (2)
- measurements (2)
- metabolic modeling (2)
- mobile networks (2)
- multipath (2)
- multipath scheduling (2)
- natural variation (2)
- navigation (2)
- network calculus (2)
- networks (2)
- ontology (2)
- perception (2)
- prediction (2)
- regular languages (2)
- reguläre Sprachen (2)
- satellite communication (2)
- scalability (2)
- segmentation (2)
- self-aware computing (2)
- sensor (2)
- smart speaker (2)
- spatial presence (2)
- tinnitus (2)
- virtual agent (2)
- virtual environments (2)
- 26S RDNA Data (1)
- 3D Laser Scanning (1)
- 3D Pointcloud (1)
- 3D Punktwolke (1)
- 3D Reconstruction (1)
- 3D Sensor (1)
- 3D Vision (1)
- 3D collation (1)
- 3D fluoroscopy (1)
- 3D point cloud (1)
- 3D thermal mapping (1)
- 3D viewer (1)
- 3D-Rekonstruktion (1)
- 3D-reconstruction methods (1)
- 3DTK toolkit (1)
- 4D-GIS (1)
- 4G Networks (1)
- 5G core network (1)
- 5G-ATSSS (1)
- 5GC (1)
- 6DOF Pose Estimation (1)
- 6G (1)
- ACKR4 (1)
- AI (1)
- AKT (1)
- ATSSSS (1)
- AVA (1)
- Abhängigskeitsgraph (1)
- Accessibility (1)
- Add-on-Miss (1)
- Admission Control (1)
- Analysis (1)
- Anwendungsfall (1)
- Approximation (1)
- Arctic (1)
- Arterie (1)
- Artery (1)
- Automat <Automatentheorie> (1)
- Automata Theory (1)
- Automatentheorie (1)
- Autonomous Robot (1)
- Autonomous UAV (1)
- Autonomous multi-vehicle systems (1)
- Autorotation (1)
- BPM (1)
- BPMN (1)
- Backbone-Netz (1)
- Background Knowledge (1)
- Balloon (1)
- Banks Islands (1)
- Barcodes (1)
- Bayes-Klassifikator (1)
- Benutzererlebnis (1)
- Benutzerforschung (1)
- Benutzerinteraktion (1)
- Berechenbarkeit (1)
- Berechnungskomplexität (1)
- Betriebssystem (1)
- Bewegungskompensation (1)
- Bewegungskoordination (1)
- Bildverarbeitung (1)
- Biological Networks (1)
- Biology (1)
- BitTorrent (1)
- Bodenstation (1)
- Boolean Grammar (1)
- Boolean equivalence (1)
- Boolean function (1)
- Boolean functions (1)
- Boolean hierarchy (1)
- Boolean isomorphism (1)
- Boolean tree (1)
- Boolesche Funktionen (1)
- Boolesche Grammatik (1)
- Boolesche Hierarchie (1)
- CASE (1)
- CD4+T cells (1)
- CD8+T cells (1)
- CD95 (1)
- CETCH cycle (1)
- CHI Conference (1)
- CLIP (1)
- CO2-sequestration (1)
- COVID-19 (1)
- Caenorhabditis elegans (1)
- Call Graph (1)
- Causes of revelation (1)
- Chapters arrangement (1)
- Charged aerosol detector (CAD) (1)
- Chord (1)
- Chronology of revelation (1)
- Clones (1)
- Cloud Gaming (1)
- Clustering (1)
- Colonial volvocales chlorophyta (1)
- Communication Networks (1)
- Complex Systems (1)
- Complexity Theory (1)
- Compression (1)
- Computational Geometry (1)
- Computational complexity (1)
- Computer software (1)
- Computersimulation (1)
- Computerspiel (1)
- Computerunterstütztes Lernen (1)
- Containerization (1)
- Content Distribution (1)
- Crowdsensing (1)
- CubeSat (1)
- DHT (1)
- DNA (1)
- DNA storage (1)
- Daedalus-Projekt (1)
- Dasycladales chlorophyta (1)
- Datenkommunikationsnetz (1)
- Deep Georeferencing (1)
- Deep Learning (1)
- Dependency Graph (1)
- Design (1)
- Design and Development (1)
- Dezentrale Regelung (1)
- Dichotomy (1)
- Digital Elevation Model (1)
- Dijkstra’s algorithm (1)
- Diskrete Simulation (1)
- Distributed Space Systems (1)
- Domänenspezifische Sprache (1)
- Dot-Depth-Hierarchie (1)
- Drahtloses vermaschtes Netz (1)
- Dreidimensionale Rekonstruktion (1)
- Dreieck (1)
- Dynamic Memory Management (1)
- Dynamische Speicherverwaltung (1)
- E8 symmetry (1)
- EEG (1)
- EEG frequency band analysis (1)
- EEG preprocessing (1)
- EEG processing (1)
- EPM (1)
- Echtzeitsystem (1)
- Echzeit (1)
- Edge-MEC-Cloud (1)
- Edge-based Intelligence (1)
- Eingebettetes System (1)
- Elasticity tensor (1)
- Elastizitätstensor (1)
- Embedded Systems (1)
- Emotion inference (1)
- Emotionserkennung (1)
- Emotionsinterpretation (1)
- Endpoint Mobility (1)
- Energieeffizienz (1)
- Energy efficiency (1)
- Erfüllbarkeitsproblem (1)
- Erweiterte Realität (1)
- Euclidean plane (1)
- Euklidische Ebene (1)
- Expert System (1)
- Expertensystem (1)
- FIFO caching strategies (1)
- FLIMbee (1)
- Fairness (1)
- Fatty acids (1)
- Feature Based Registration (1)
- Fehlertoleranz (1)
- Fernsteuerung (1)
- Flugkörper (1)
- Forces (1)
- Formale Sprache (1)
- Formation (1)
- Formation Flight (1)
- Formationsbewegung (1)
- Formmessung (1)
- Fragmentation (1)
- Fragmentierung (1)
- Frames (1)
- Funkressourcenverwaltung (1)
- GNSS/INS integrated navigation (1)
- Gastroenterologische Endoskopie (1)
- Gefühl (1)
- Generalisierung <Kartografie> (1)
- Generation Problem (1)
- Generierungsproblem (1)
- Genetic Optimization (1)
- Genetische Optimierung (1)
- Georeferenzierung (1)
- Gothenburg Modell (1)
- Gothenburg model of collation process (1)
- Gradient boosted trees (GBT) (1)
- Graphentheorie (1)
- Ground Station Networks (1)
- H.264 SVC (1)
- H.264/SVC (1)
- HGPS (1)
- HHblits (1)
- HMD (Head-Mounted Display) (1)
- HSPA (1)
- HTML (1)
- HTTP adaptive video streaming (1)
- Halbordnungen (1)
- Hardware (1)
- Herzkatheter (1)
- Herzkathetereingriff (1)
- High-performance liquid chromatography (HPLC) (1)
- Hintergrundwissen (1)
- Historical Maps (1)
- Historische Karte (1)
- Historische Landkarten (1)
- Hittitology (1)
- Human-centered computing / Access (1)
- Human-centered computing / Human computer interaction (HCI) / Interaction paradigms / Mixed / augmented reality (1)
- Human-centered computing / Human computer interaction (HCI) / Interaction paradigms / Virtual reality (1)
- Human-centered computing / Human computer interaction (HCI) / Interactiondevices (1)
- Human-centered computing / Human computerinteraction (HCI) / Interaction techniques (1)
- Hurwitz theorem (1)
- I-tasser (1)
- ICEP (1)
- IEEE 802.11e (1)
- IEEE 802.15.4 (1)
- IGFBP2 (1)
- III secretion (1)
- INS/LIDAR integrated navigation (1)
- IP (1)
- IT security (1)
- Image Aesthetic Assessment (1)
- Image Registration (1)
- ImageJ (1)
- Implementierung <Informatik> (1)
- Industrial internet (1)
- Industrie 4.0 (1)
- Information Extraction (1)
- Information Visualization (1)
- Instrument Control Toolbox (1)
- Intelligent Virtual Agents (1)
- InteractionSuitcase (1)
- Interaktion (1)
- Internet (1)
- Internet Protokoll (1)
- IoT-driven processes (1)
- IronChip Evaluation Package (1)
- Isomorphie (1)
- Itinerare (1)
- Itineraries (1)
- JCAS (1)
- JSF (1)
- Jacobian matrix (1)
- Java 3D (1)
- Java <Programmiersprache> (1)
- Java Frameworks (1)
- Java Message Service (1)
- Kademlia (1)
- Kanalzugriff (1)
- Karte (1)
- Kathará (1)
- Kerneldensity estimation (1)
- Klassendiagramm (1)
- Klassifikation (1)
- Klima (1)
- Knowledge Discovery (1)
- Knowledge Management System (1)
- Knowledge Modeling (1)
- Knowledge representation (1)
- Knowledge-based System (1)
- Knowledge-based Systems Engineering (1)
- Kombinatorik (1)
- Kommunikation (1)
- Kommunikationsnetze (1)
- Komplexes System (1)
- Komplexitätsklasse (1)
- Komplexitätsklasse NP (1)
- Konvexe Zeichnungen (1)
- Kooperierende mobile Roboter (1)
- Kreuzungsminimierung (1)
- Kryoelektronenmikroskopie (1)
- Kurve (1)
- Künstliche Intelligenz (1)
- LC-MS/MS (1)
- LFU (1)
- LRU (1)
- Land Cover Classification (1)
- Land plants (1)
- Landkartenbeschriftung (1)
- Landnutzungskartierung (1)
- Laser scanning (1)
- Lava (1)
- Lawhul-Mahfuz (1)
- Lee Smolin (1)
- Lidar (1)
- Lifetime spectroscopy (1)
- Lightning (1)
- Link rate adaptation (1)
- Linkratenanpassung (1)
- Linux (1)
- LoRa (1)
- LoRaWAN (1)
- LoRaWan (1)
- Logic Programming (1)
- Logische Programmierung (1)
- Lunar Caves (1)
- Lunar Exploration (1)
- MAC (1)
- MDR (1)
- MTC (1)
- MVC <Software> (1)
- Mackenzie-River-Delta (1)
- Mapping (1)
- Mars (1)
- Mashup (1)
- Mashup <Internet> (1)
- Mathematische Modellierung (1)
- Mathematisches Modell (1)
- Measurement (1)
- Medical Image Analysis (1)
- Medium <Physik> (1)
- Medizin (1)
- Mehrfahrzeugsysteme (1)
- Mehrkriterielle Optimierung (1)
- Mehrpfadübertragung (1)
- Mehrschichtnetze (1)
- Mehrschichtsystem (1)
- Mensch-Maschine-Kommunikation (1)
- Mesh Networks (1)
- Mesh Netze (1)
- Metaverse (1)
- Methode (1)
- Microarray (1)
- Middleware (1)
- Miniaturisierung (1)
- Minimally invasive vascular intervention (1)
- Mobiles Internet (1)
- Mobilfunk (1)
- Modell (1)
- Modellierungstechniken (1)
- Modularität (1)
- Molecular systematics (1)
- Mond (1)
- Multi-Layer (1)
- Multi-Network Service (1)
- Multi-Netzwerk Dienste (1)
- Multi-Paradigm Programming (1)
- Multi-Paradigm Programming Framework (1)
- Multipath Transmission (1)
- Mustererkennung (1)
- Mycoplasma (1)
- Mycoplasma pneumoniae (1)
- NP (1)
- NP-Vollständigkeit (1)
- NP-complete sets (1)
- NP-hartes Problem (1)
- NP-schweres Problem (1)
- Naïve Bayesian (1)
- Network Emulator (1)
- Network Experiments (1)
- Network Management (1)
- Network Measurements (1)
- Network Virtualization (1)
- Networks (1)
- Netzplantechnik (1)
- Netzplanung (1)
- Netzvirtualisierung (1)
- Netzwerkplanung (1)
- Netzwerkvirtualisierung (1)
- Neuromuscular junctions (1)
- Neuronales Netz (1)
- Newton Methods (1)
- Newton-Verfahren (1)
- Next Generation Networks (1)
- Nichtholonome Fahrzeuge (1)
- Nichtlineare Regelung (1)
- Nuclear RDNA (1)
- Object Detection (1)
- Object-Oriented Programming (1)
- Objektorientierte Programmierung (1)
- Operator (1)
- Optical Flow (1)
- Optimale Kontrolle (1)
- Optimierungsproblem (1)
- Optimization on Lie Groups (1)
- Overlapping (1)
- Overlay (1)
- Overlay Netzwerke (1)
- Overlay networks (1)
- Overlays (1)
- P4-INT (1)
- Panorama Images (1)
- Partition <Mengenlehre> (1)
- Partitionen (1)
- Path Computation Element (1)
- Pattern Recognition (1)
- Peer-to-Peer (1)
- Performance Analysis (1)
- Performance Enhancing Proxies (1)
- Performance Management (1)
- Performance Modeling (1)
- Pfadberechnungselement (1)
- Picosatellite (1)
- Place of revelation (1)
- Planare Graphen (1)
- Planung (1)
- Poisson surface reconstruction (1)
- PolSAR (1)
- Polyeder (1)
- Polypektomie (1)
- Positron annihilation spectroscopy (1)
- Post's Classes (1)
- Postsche Klassen (1)
- Prediction (1)
- Processing Model (1)
- Processing model (1)
- Profile distances (1)
- Programmierbare logische Anordnung (1)
- Publish-Subscribe-System (1)
- QUIC (1)
- QoS (1)
- Quality of Experience (QoE) (1)
- Quality of Experience QoE (1)
- Quality of Service (1)
- Quality of Service (QoS) (1)
- Quality-of-Experience (1)
- Quality-of-Service (1)
- Quality-of-Service (QoS) (1)
- Quantitative structure-property relationship modeling (QSPR) (1)
- Quantor (1)
- RBCL Gene-sequences (1)
- RGB-D (1)
- RNA sequencing (1)
- RRM (1)
- Radarfernerkundung (1)
- Raumdaten (1)
- Real-Time Operating Systems (1)
- Real-time (1)
- Reconstruction of original text (1)
- Refactoring (1)
- Reference Architecture (1)
- Regelung (1)
- Registration (1)
- Registrierung (1)
- Registrierung <Bildverarbeitung> (1)
- Reguläre Sprache (1)
- Relief <Geografie> (1)
- Rendezvous (1)
- Reproducibility (1)
- Resource and Performance Management (1)
- Ressourcen Management (1)
- Ressourcenallokation (1)
- Rettungsroboter (1)
- SARS-CoV-2 (1)
- SBA (1)
- SDN/NVF (1)
- SMLM (1)
- SNP (1)
- Satellit (1)
- Scatter Plot (1)
- Scheduling (1)
- Search-and-Rescue (1)
- Secondary structure (1)
- Selbstkalibrierung (1)
- Self-calibration (1)
- Semantic Web (1)
- Semantics (1)
- Semantik (1)
- Sensing-aaS (1)
- Sensor (1)
- Septins (1)
- Service Mobility (1)
- Sichtbarkeit (1)
- Similarity Measure (1)
- Simulator (1)
- Situationsbewusstsein (1)
- Skype (1)
- Small Satellites (1)
- Smart User Interaction (1)
- Social Web (1)
- Software (1)
- Software Performance Engineering (1)
- Software Performance Modeling (1)
- Software architecture (1)
- Software design (1)
- Software product lines (1)
- Source Code Visualization (1)
- Soziale Software (1)
- Spherical Robot (1)
- Spring (1)
- Stages of Prophet Mohammad’s messengership (1)
- Standardisierung (1)
- Standortproblem (1)
- Statistical classifiers (1)
- Statistics (1)
- Statistische Mechanik (1)
- Statistische Physik (1)
- Sternfreie Sprache (1)
- Steuerung (1)
- Stiffness (1)
- Stochastic Algorithms (1)
- Stochastische Optimierung (1)
- Strahlentherapie (1)
- Straubing-Th´erien-Hierarchie (1)
- Structure-from-Motion (1)
- Strukturelle Komplexität (1)
- Struts (1)
- Subgroup Mining (1)
- Subgruppenentdeckung (1)
- Substruktur (1)
- Support Vector Machine (1)
- Synapses (1)
- Synaptic vesicles (1)
- Synthetic Aperture Radar (1)
- TSN (1)
- TTL (1)
- TTL validation of data consistency (1)
- Telematik (1)
- Terramechanics (1)
- Testbed (1)
- Text categorization (1)
- Text segmentation (1)
- Theoretical computer science (1)
- Thermografie (1)
- Time resolved measurements (1)
- Tomografie (1)
- Topografie (1)
- Torque (1)
- Trainingssystem (1)
- Travelling-salesman-Problem (1)
- Tumor motion (1)
- Tumorbewegung (1)
- U-Bahnlinienplan (1)
- UI and Interaction Design (1)
- UML Klassendiagramm (1)
- UML class diagram (1)
- UMTS (1)
- URL (1)
- URLLC (1)
- Underwater Mapping (1)
- Underwater Scanning (1)
- Unmanned Aerial Vehicle (1)
- Unstetige Regelung (1)
- Usability (1)
- Use case (1)
- User Behavior (1)
- User Participation (1)
- V-antigen (1)
- VNF (1)
- VPN (1)
- Variability (1)
- Venus (1)
- Veranstaltung (1)
- Verbotenes Muster (1)
- Verbände (1)
- Verkehrslenkung (1)
- Verteilung von Inhalten (1)
- Vesicles (1)
- Video Quality Monitoring (1)
- Video Streaming (1)
- Videoübertragung (1)
- Virtualisierung (1)
- Virtuelles Netzwerk (1)
- Visibility (1)
- Visual Text Mining (1)
- Visual Tracking (1)
- Visualized Kathará (1)
- Voice-over-IP (VoIP) (1)
- Vorhersage (1)
- WH2 domain (1)
- WLAN (1)
- WNT (1)
- Warteschlangentheorie (1)
- Web service (1)
- WebGL (1)
- Webservice Composition (1)
- WhatsApp (1)
- Wheel (1)
- Winkel (1)
- Wire relaxation (1)
- Wireless LAN (1)
- Wireless Sensor/Actuator Systems (1)
- Wissensbanksystem (1)
- Wissensbasiertes System (1)
- Wissenschaftliche Beobachtung (1)
- Wissensendeckung (1)
- Worterweiterungen (1)
- Wrapper (1)
- Wrappers (1)
- XML model (1)
- XR-artificial intelligence combination (1)
- XR-artificial intelligence continuum (1)
- Yersinia enterocolitica (1)
- Yolk protein (1)
- YouTube (1)
- Zebrafish (1)
- Zeichnen von Graphen (1)
- Zeitdiskretes System (1)
- Zugangskontrolle (1)
- Zählprobleme (1)
- abgeschlossene Klassen (1)
- acrophobia (1)
- actin nucleation (1)
- adaptation (1)
- adaptation models (1)
- administrative boundary (1)
- admission control (1)
- adult learning (1)
- advertising effectiveness (1)
- aerodynamics (1)
- aerospace (1)
- agent-based models (1)
- agents (1)
- agile Prozesse (1)
- agile processes (1)
- aging (1)
- alignment (1)
- anamnesis tool (1)
- aneurysm (1)
- angular schematization (1)
- annotation (1)
- anomaly detection (1)
- anomaly prediction (1)
- ant-colony optimization (1)
- anthropomorphism (1)
- anxiety (1)
- apixaban (1)
- application design (1)
- approximation algorithms (1)
- arabidopsis thaliana (1)
- arabidpsis thaliana (1)
- architectural design (1)
- arithmetic calculations (1)
- artificial intelligence education (1)
- artificial intelligence literacy (1)
- augmentation (1)
- automatic Layout (1)
- automatisches Layout (1)
- autonomic orchestration (1)
- autonomous (1)
- autonomous UAV (1)
- autorotation (1)
- availability (1)
- avatar embodiment (1)
- avatars (1)
- background knowledge (1)
- baseline detection (1)
- behavior (1)
- behavior change (1)
- behavior perception (1)
- bibliometric analysis (1)
- binary decision diagram (1)
- binary tanglegram (1)
- biofuel (1)
- biohybrid systems (1)
- bioinformatics (1)
- biological development (1)
- biomanufacturing (1)
- biosignals (1)
- bit (1)
- boundary labeling (1)
- brain (1)
- building (1)
- caenorhabditis elegans (1)
- camera orientation (1)
- car-like robots (1)
- carbon (1)
- carboxylation (1)
- cardiac magnetic resonance (1)
- caspase-3 (1)
- cell membranes (1)
- cerebral ischemia (1)
- certifying algorithm (1)
- chain cover (1)
- channel management (1)
- cisplatin (1)
- classification (1)
- climate (1)
- cloud-native (1)
- co-authorships (1)
- co-inventorships (1)
- cognitive impairment (1)
- coherence (1)
- collaboration (1)
- collision (1)
- colony-stimulating factor (1)
- combination therapy (1)
- communication models (1)
- communication networks (1)
- community detection (1)
- comparative sequence analysis (1)
- competitive location (1)
- complex traits (1)
- complexity (1)
- compressed sensing (1)
- computational (1)
- computational complexity (1)
- computer performance evaluation (1)
- computergestützte Softwaretechnik (1)
- computers as social actors (1)
- condition prediction (1)
- congruence (1)
- connector (1)
- constrained forest (1)
- contact representation (1)
- container virtualization (1)
- content-based image retrieval (1)
- continuous-time SLAM (1)
- conversational agent (1)
- conversational agents (1)
- convex bipartite graph (1)
- corticotropin-releasing hormone (1)
- cost-sensitive learning (1)
- counting problems (1)
- crosstalk (1)
- crowdsourced QoE measurements (1)
- crowdsourced measurements (1)
- crowdsourced network measurements (1)
- cryo-EM (1)
- cryo-ET (1)
- crystal growth (1)
- crystallization (1)
- cultural and media studies (1)
- culturally aware (1)
- cuneiform (1)
- curves (1)
- cyber-physical systems (1)
- cybersickness (1)
- cytokine profiling (1)
- d3web.Train (1)
- dSTORM (1)
- data plane programming (1)
- data stream processing (1)
- data structure (1)
- data warehouse (1)
- dataplane programming (1)
- decision support system (1)
- decision-making (1)
- decoding error rate (1)
- deep metric learning (1)
- definite clause grammars (1)
- deformation-based method (1)
- delay QoS exponent (1)
- delay bound violation probability (1)
- delay constrained (1)
- dementia (1)
- descent (1)
- design cycle (1)
- detection time simulation (1)
- diagnostic accuracy (1)
- dial a ride (1)
- differentiation (1)
- digital twin (1)
- dimensions of proximity (1)
- direct oral anticoagulants (1)
- direct thrombin inhibitor (1)
- discrete-time analysis (1)
- disease (1)
- disjoint multi-paths (1)
- disruption project (1)
- distance-based classifier (1)
- distributed control (1)
- docker (1)
- drug (1)
- drug-minded protein (1)
- dynamic flow migration (1)
- dynamic programming (1)
- eHealth (1)
- eco-metabolomics (1)
- ecological momentary assessment (1)
- edge labeled graphs (1)
- effective Bandwidth (1)
- efficient algorithm (1)
- electroencephalography (1)
- electrolytes (1)
- electronic health records (1)
- elementary mode analysis (1)
- elementary modes (1)
- elevated plus-maze (1)
- embedding techniques (1)
- emergent time (1)
- emotions (1)
- empathy (1)
- emulation (1)
- encryption (1)
- endurance (1)
- energy efficiency (1)
- environmental sound (1)
- enzyme (1)
- event detection (1)
- event-related potentials-ERP (1)
- exercise intensity (1)
- experience (1)
- experimental evaluation (1)
- expertise framing (Min5-Max 8) (1)
- expression (1)
- expression signature (1)
- extended reality (1)
- extended reality (XR) (1)
- factor XA inhibitor (1)
- failure prediction (1)
- fast reroute (1)
- fault detection (1)
- feature matching (1)
- federated learning (1)
- few-shot learning (1)
- fixed-parameter tractability (1)
- flies (1)
- fluoroscopy (1)
- food quality (1)
- force dynamics (1)
- foreign language learning and teaching (1)
- formation driving (1)
- formation flight (1)
- fruit temperature (1)
- functional analysis (1)
- future Internet architecture (1)
- future energy grid exploration (1)
- game mechanics (1)
- gamification (1)
- gamma (1)
- generative systems (1)
- genes (1)
- genetic algorithm (1)
- genetic regulatory network (1)
- geospatial data (1)
- global IPX network (1)
- graph (1)
- graph algorithm (1)
- graph decomposition (1)
- graphs (1)
- green systems biology (1)
- group-based communication (1)
- handwriting (1)
- haptic data (1)
- hardness (1)
- hardware-in-the-loop simulation (1)
- hardware-in-the-loop streaming system (1)
- hepatotoxicity (1)
- heuristics (1)
- hierarchy (1)
- histidine kinase (1)
- historical images (1)
- hit ratio analysis and simulation (1)
- homology modeling (1)
- hospital data (1)
- human body weight (1)
- human computer interaction (HCI) (1)
- human-artificial intelligence interaction (1)
- human-artificial intelligence interface (1)
- human-centered design (1)
- human-centered, human-robot (1)
- humantechnology interaction (1)
- human–computer interaction (1)
- hybrid access (1)
- hypotonic (1)
- hypotonic solutions (1)
- illusion of self-motion (1)
- image classification (1)
- image processing (1)
- image schemas (1)
- imbalanced regression (1)
- immersive advertising (1)
- immersive classroom (1)
- immersive classroom management (1)
- immersive learning technologies (1)
- immunity (1)
- implicit association test (1)
- in situ analysis (1)
- independent crossing (1)
- induced matching (1)
- inflation (1)
- informal education (1)
- information extraction (1)
- information systems and information technology (1)
- inhibitor (1)
- intelligent transportation systems (1)
- intelligent vehicles (1)
- intelligent virtual agents (1)
- intelligent voice assistant (1)
- intelligente Applikationen (1)
- intention-behavior-gap (1)
- inter-coder reliability (1)
- interaction (1)
- interactive authoring system (1)
- interactive collation of textual variants (1)
- intercultural learning and teaching (1)
- interdisciplinary education (1)
- intermediate host (1)
- internal transcribed spacer 2 (1)
- internet protocol (1)
- internet traffic (1)
- interpolation (1)
- intervention design (1)
- intervention evaluation (1)
- intraoperative imaging (1)
- invasive vascular interventions (1)
- iowa gambling task (1)
- isotonic (1)
- key-insight extraction (1)
- kinect (1)
- labeling (1)
- land-cover area (1)
- landing (1)
- language-image pre-training (1)
- latency (1)
- lattices (1)
- layout recognition (1)
- learning environments (1)
- least cost (1)
- life-span regulation (1)
- lifetime spectroscopy (1)
- light-gated proteins (1)
- load balancing (1)
- local energy system (1)
- locomotion (1)
- logic programming (1)
- logistics (1)
- long-term analysis (1)
- lymphotoxicity (1)
- malaria (1)
- map projections (1)
- mapping (1)
- markers (1)
- mathematical model (1)
- measurement (1)
- media analysis (1)
- media equation (1)
- medical analytics (1)
- medical device regulation (1)
- medical device software (1)
- medical records (1)
- medieval manuscripts (1)
- meditation (1)
- membrane protein (1)
- membrane proteins (1)
- memory immune responses (1)
- metabolic flux (1)
- metabolism (1)
- metabolomics (1)
- metastasis (1)
- methylene blue (1)
- metro map (1)
- mice (1)
- microbes (1)
- mindfulness (1)
- misconceptions (1)
- mixed reality (1)
- mixed-cultural (1)
- mixed-cultural settings (1)
- mobile application (1)
- mobile instant messaging (1)
- mobile messaging application (1)
- mobile robots (1)
- mobile streaming (1)
- model following (1)
- model output statistics (1)
- model predictive control (1)
- model-based diagnosis (1)
- modeling techniques (1)
- modules (1)
- molecular systematics (1)
- monotone drawing (1)
- morphing (1)
- mouse (1)
- multi-vehicle formations (1)
- multi-vehicle rendezvous (1)
- multimodal fusion (1)
- multimodal interface (1)
- multimodal learning (1)
- multipath communication (1)
- multipath packet scheduling (1)
- multiple myeloma (1)
- multirotors (1)
- multiscale encoder (1)
- mutation (1)
- mycoplasma (1)
- nano-satellite (1)
- nanocellulose (1)
- natural interfaces (1)
- natural language processing (1)
- natural language processing · · · (1)
- natural user interfaces (1)
- network (1)
- network design (1)
- network planning (1)
- network softwarization (1)
- network upgrade (1)
- network virtualization (1)
- networked robotics (1)
- networking (1)
- neume notation (1)
- neural architecture (1)
- noise measurement (1)
- non-native accent (1)
- non-terrestrial networks (1)
- nonholonomic vehicles (1)
- nonhuman-primates (1)
- nonverbal behavior (1)
- object detection (1)
- octree (1)
- omics (1)
- optical music recognition (1)
- optimization (1)
- orchestration (1)
- organogenesis (1)
- origin (1)
- overprovisioning (1)
- oxidative stress (1)
- packet reception method (1)
- painful (1)
- pangolin (1)
- particle picking (1)
- partitions (1)
- passage of time (1)
- passive haptic feedback (1)
- path computation (1)
- pathway (1)
- performance (1)
- performance analysis (1)
- performance evaluation (1)
- performance liquid chromatography (1)
- performance monitoring (1)
- performance prediction (1)
- permeability (1)
- pestis infection (1)
- phase space (1)
- phase transition (1)
- photorespiration (1)
- phylogenetic tree (1)
- phylogeny (1)
- place-illusion (1)
- plain orchestrating service (1)
- plausibility (1)
- plausibility-illusion (1)
- pneumoniae (1)
- pneumonic plague (1)
- point cloud (1)
- point cloud compression (1)
- point-to-plane measure (1)
- point-to-point measure (1)
- pollution (1)
- pos (1)
- posets (1)
- positioning (1)
- precision horticulture (1)
- precision training (1)
- presence (1)
- private chat groups (1)
- procedural content generation (1)
- procedural fusion methods (1)
- processing pipeline (1)
- progeria (1)
- promoter (1)
- prompt engineering (1)
- protein (1)
- protein chip (1)
- protein-interaction networks (1)
- pseudomas-syringae (1)
- psychomotor training (1)
- psychophyisology (1)
- public speaking (1)
- pulse simulation (1)
- quadcopter (1)
- quadcopters (1)
- quality assurance (1)
- quality evaluation (1)
- quality of experience (1)
- quality of experience prediction (1)
- quantification (1)
- qubit (1)
- radio resource management (1)
- radiology (1)
- ransomware (1)
- real world evidence (1)
- real-world application (1)
- realism (1)
- receding horizon control (1)
- receptor (1)
- recombinant protein rVE (1)
- recombination (1)
- recommender system (1)
- reconfiguration (1)
- reload cost (1)
- remote control (1)
- research methods (1)
- resilience (1)
- response regulator (1)
- ribosomal RNA (1)
- rich vehicle routing problem (1)
- richtersius coronifer (1)
- right angle crossing (1)
- robotics (1)
- robustness (1)
- rotorcraft (1)
- rotors (1)
- routing (1)
- sample weighting (1)
- satisfiability problems (1)
- scalability evaluation (1)
- scalable quadcopter (1)
- scheduling (1)
- science, technology and society (1)
- secondary structure (1)
- secure group communication (1)
- self-adaptive (1)
- self-adaptive systems (1)
- self-assembly (1)
- self-aware computing systems (1)
- self-managing systems (1)
- semantic fusion (1)
- semantic understanding (1)
- semantic web (1)
- semantical aesthetic (1)
- semantische Ästhetik (1)
- sensitivity analysis (1)
- sensor devices (1)
- sensor fusion (1)
- sensor networks (1)
- sentinel (1)
- sequence alignment (1)
- serious games (1)
- serum (1)
- service based software architecture (1)
- service brokerage (1)
- service-curve estimation (1)
- sesnsors (1)
- set (1)
- shootin-1 (1)
- short block-length (1)
- shortest path routing (1)
- signal processing (1)
- signaling traffic (1)
- simulation system (1)
- simultaneous embedding (1)
- single-electron transistors (1)
- sketching (1)
- smart meter data utilization (1)
- smooth orthogonal drawing (1)
- snow shoveling (1)
- social VR (1)
- social interaction (1)
- social relationship (1)
- social robot (1)
- social robotics (1)
- social role (1)
- socially interactive agents (1)
- software defined network (1)
- spanning tree (1)
- spire (1)
- stability (1)
- stable state (1)
- standardization (1)
- state management (1)
- statistical validity (1)
- statistics and numerical data (1)
- stereotypes (1)
- stream processing (1)
- stroke (1)
- structural complexity (1)
- student simulation (1)
- stylus (1)
- sun exposure (1)
- sunburn (1)
- superoxide-dismutase (1)
- supervised learning (1)
- surface model (1)
- survey (1)
- survival (1)
- sustainability (1)
- switching navigation (1)
- synthetic biology (1)
- synthetic pathways (1)
- system architecture design (1)
- systematic literature review (1)
- systematic review (1)
- table extraction (1)
- table understanding (1)
- taxonomy (1)
- teacher education (1)
- technology-supported learning (1)
- telematics (1)
- temperature (1)
- text categorization (1)
- text line detection (1)
- text supervision (1)
- theory (1)
- therapeutic application (1)
- thermal camera (1)
- thermal point cloud (1)
- time calibration (1)
- time perception (1)
- time series (1)
- timestamping method (1)
- tolerance (1)
- tomography (1)
- tonicity (1)
- tools (1)
- topology (1)
- training systems (1)
- trait anxiety (1)
- trajectory planning (1)
- transcription (1)
- transformations (1)
- transformer (1)
- translational neuroscience (1)
- transmission (1)
- transport microenvironments (1)
- transport protocols (1)
- transportation (1)
- tree (1)
- trust (1)
- trustworthiness (1)
- unmanned aerial vehicle (1)
- unmanned aerial vehicles (1)
- usability evaluation (1)
- use cases (1)
- user experience (1)
- user interaction (1)
- user interfaces (1)
- user study (1)
- user-generated content (1)
- v (1)
- vaccine (1)
- validation (1)
- vection (1)
- vehicle dynamics (1)
- vehicular navigation (1)
- verbal behaviour (1)
- vernetzte Roboter (1)
- virtual agent interaction (1)
- virtual audience (1)
- virtual humans (1)
- virtual queue (1)
- virtual reality training (1)
- virtual stimuli (1)
- virtual tunnel (1)
- virtual-reality-continuum (1)
- visual analytics (1)
- visual proteomics (1)
- vitellogenin (1)
- voice assistant (1)
- voice-based artificial intelligence (1)
- vom Nutzer erfahrene Dienstgüte QoE (1)
- voting location (1)
- water stress (1)
- waypoint parameter (1)
- wearable (1)
- wheel (1)
- word clouds (1)
- word extensions (1)
- zooming (1)
- zukünftige Kommunikationsnetze (1)
- zukünftiges Internet (1)
- Ähnlichkeitsmaß (1)
- Überlappung (1)
Institute
- Institut für Informatik (183)
- Theodor-Boveri-Institut für Biowissenschaften (29)
- Institut Mensch - Computer - Medien (17)
- Institut für deutsche Philologie (17)
- Institut für Klinische Epidemiologie und Biometrie (7)
- Center for Computational and Theoretical Biology (4)
- Graduate School of Science and Technology (3)
- Medizinische Klinik und Poliklinik II (3)
- Institut für Funktionsmaterialien und Biofabrikation (2)
- Institut für Geographie und Geologie (2)
Schriftenreihe
Sonstige beteiligte Institutionen
- Cologne Game Lab (2)
- Birmingham City University (1)
- DATE Lab, KITE Research Insititute, University Health Network, Toronto, Canada (1)
- EMBL Heidelberg (1)
- INAF Padova, Italy (1)
- Jacobs University Bremen, Germany (1)
- Open University of the Netherlands (1)
- Servicezentrum Medizin-Informatik (Universitätsklinikum) (1)
- Social and Technological Systems (SaTS) lab, School of Art, Media, Performance and Design, York University, Toronto, Canada (1)
- TH Köln (1)
In this doctoral thesis we cover the performance evaluation of next generation data plane architectures, comprised of complex software as well as programmable hardware components that allow fine granular configuration. In the scope of the thesis we propose mechanisms to monitor the performance of singular components and model key performance indicators of software based packet processing solutions. We present novel approaches towards network abstraction that allow the integration of heterogeneous data plane technologies into a singular network while maintaining total transparency between control and data plane. Finally, we investigate a full, complex system consisting of multiple software-based solutions and perform a detailed performance analysis. We employ simulative approaches to investigate overload control mechanisms that allow efficient operation under adversary conditions. The contributions of this work build the foundation for future research in the areas of network softwarization and network function virtualization.
Starfree regular languages can be build up from alphabet letters by using only Boolean operations and concatenation. The complexity of these languages can be measured with the so-called dot-depth. This measure leads to concatenation hierarchies like the dot-depth hierarchy (DDH) and the closely related Straubing-Thérien hierarchy (STH). The question whether the single levels of these hierarchies are decidable is still open and is known as the dot-depth problem. In this thesis we prove/reprove the decidability of some lower levels of both hierarchies. More precisely, we characterize these levels in terms of patterns in finite automata (subgraphs in the transition graph) that are not allowed. Therefore, such characterizations are called forbidden-pattern characterizations. The main results of the thesis are as follows: forbidden-pattern characterization for level 3/2 of the DDH (this implies the decidability of this level) decidability of the Boolean hierarchy over level 1/2 of the DDH definition of decidable hierarchies having close relations to the DDH and STH Moreover, we prove/reprove the decidability of the levels 1/2 and 3/2 of both hierarchies in terms of forbidden-pattern characterizations. We show the decidability of the Boolean hierarchies over level 1/2 of the DDH and over level 1/2 of the STH. A technique which uses word extensions plays the central role in the proofs of these results. With this technique it is possible to treat the levels 1/2 and 3/2 of both hierarchies in a uniform way. Furthermore, it can be used to prove the decidability of the mentioned Boolean hierarchies. Among other things we provide a combinatorial tool that allows to partition words of arbitrary length into factors of bounded length such that every second factor u leads to a loop with label u in a given finite automaton.
This article presents a novel method for controlling a virtual audience system (VAS) in Virtual Reality (VR) application, called STAGE, which has been originally designed for supervised public speaking training in university seminars dedicated to the preparation and delivery of scientific talks. We are interested in creating pedagogical narratives: narratives encompass affective phenomenon and rather than organizing events changing the course of a training scenario, pedagogical plans using our system focus on organizing the affects it arouses for the trainees. Efficiently controlling a virtual audience towards a specific training objective while evaluating the speaker’s performance presents a challenge for a seminar instructor: the high level of cognitive and physical demands required to be able to control the virtual audience, whilst evaluating speaker’s performance, adjusting and allowing it to quickly react to the user’s behaviors and interactions. It is indeed a critical limitation of a number of existing systems that they rely on a Wizard of Oz approach, where the tutor drives the audience in reaction to the user’s performance. We address this problem by integrating with a VAS a high-level control component for tutors, which allows using predefined audience behavior rules, defining custom ones, as well as intervening during run-time for finer control of the unfolding of the pedagogical plan. At its core, this component offers a tool to program, select, modify and monitor interactive training narratives using a high-level representation. The STAGE offers the following features: i) a high-level API to program pedagogical narratives focusing on a specific public speaking situation and training objectives, ii) an interactive visualization interface iii) computation and visualization of user metrics, iv) a semi-autonomous virtual audience composed of virtual spectators with automatic reactions to the speaker and surrounding spectators while following the pedagogical plan V) and the possibility for the instructor to embody a virtual spectator to ask questions or guide the speaker from within the Virtual Environment. We present here the design, and implementation of the tutoring system and its integration in STAGE, and discuss its reception by end-users.
In this paper, we present a virtual audience simulation system for Virtual Reality (VR). The system implements an audience perception model controlling the nonverbal behaviors of virtual spectators, such as facial expressions or postures. Groups of virtual spectators are animated by a set of nonverbal behavior rules representing a particular audience attitude (e.g., indifferent or enthusiastic). Each rule specifies a nonverbal behavior category: posture, head movement, facial expression and gaze direction as well as three parameters: type, frequency and proportion. In a first user-study, we asked participants to pretend to be a speaker in VR and then create sets of nonverbal behaviour parameters to simulate different attitudes. Participants manipulated the nonverbal behaviours of single virtual spectator to match a specific levels of engagement and opinion toward them. In a second user-study, we used these parameters to design different types of virtual audiences with our nonverbal behavior rules and evaluated their perceptions. Our results demonstrate our system’s ability to create virtual audiences with three types of different perceived attitudes: indifferent, critical, enthusiastic. The analysis of the results also lead to a set of recommendations and guidelines regarding attitudes and expressions for future design of audiences for VR therapy and training applications.
Knowledge about ransomware is important for protecting sensitive data and for participating in public debates about suitable regulation regarding its security. However, as of now, this topic has received little to no attention in most school curricula. As such, it is desirable to analyze what citizens can learn about this topic outside of formal education, e.g., from news articles. This analysis is both relevant to analyzing the public discourse about ransomware, as well as to identify what aspects of this topic should be included in the limited time available for this topic in formal education. Thus, this paper was motivated both by educational and media research. The central goal is to explore how the media reports on this topic and, additionally, to identify potential misconceptions that could stem from this reporting. To do so, we conducted an exploratory case study into the reporting of 109 media articles regarding a high-impact ransomware event: the shutdown of the Colonial Pipeline (located in the east of the USA). We analyzed how the articles introduced central terminology, what details were provided, what details were not, and what (mis-)conceptions readers might receive from them. Our results show that an introduction of the terminology and technical concepts of security is insufficient for a complete understanding of the incident. Most importantly, the articles may lead to four misconceptions about ransomware that are likely to lead to misleading conclusions about the responsibility for the incident and possible political and technical options to prevent such attacks in the future.
This document presents a networking latency measurement setup that focuses on affordability and universal applicability, and can provide sub-microsecond accuracy. It explains the prerequisites, hardware choices, and considerations to respect during measurement. In addition, it discusses the necessity for exhaustive latency measurements when dealing with high availability and low latency requirements. Preliminary results show that the accuracy is within ±0.02 μs when used with the Intel I350-T2 network adapter.
Understanding the Performance of Different Packet Reception and Timestamping Methods in Linux
(2023)
This document briefly presents some renowned packet reception techniques for network packets in Linux systems. Further, it compares their performance when measuring packet timestamps with respect to throughput and accuracy. Both software and hardware timestamps are compared, and various parameters are examined, including frame size, link speed, network interface card, and CPU load. The results indicate that hardware timestamping offers significantly better accuracy with no downsides, and that packet reception techniques that avoid system calls offer superior measurement throughput.
Failure prediction is an important aspect of self-aware computing systems. Therefore, a multitude of different approaches has been proposed in the literature over the past few years. In this work, we propose a taxonomy for organizing works focusing on the prediction of Service Level Objective (SLO) failures. Our taxonomy classifies related work along the dimensions of the prediction target (e.g., anomaly detection, performance prediction, or failure prediction), the time horizon (e.g., detection or prediction, online or offline application), and the applied modeling type (e.g., time series forecasting, machine learning, or queueing theory). The classification is derived based on a systematic mapping of relevant papers in the area. Additionally, we give an overview of different techniques in each sub-group and address remaining challenges in order to guide future research.
Packets sent over a network can either get lost or reach their destination. Protocols like TCP try to solve this problem by resending the lost packets. However, retransmissions consume a lot of time and are cumbersome for the transmission of critical data. Multipath solutions are quite common to address this reliability issue and are available on almost every layer of the ISO/OSI model. We propose a solution based on a P4 network to duplicate packets in order to send them to their destination via multiple routes. The last network hop ensures that only a single copy of the traffic is further forwarded to its destination by adopting a concept similar to Bloom filters. Besides, if fast delivery is requested we provide a P4 prototype, which randomly forwards the packets over different transmission paths. For reproducibility, we implement our approach in a container-based network emulation system called Kathará.
In network research, reproducibility of experiments is not always easy to achieve. Infrastructures are cumbersome to set up or are not available due to vendor-specific devices. Emulators try to overcome those issues to a given extent and are available in different service models. Unfortunately, the usability of emulators requires time-consuming efforts and a deep understanding of their functionality. At first, we analyze to which extent currently available open-source emulators support network configurations and how user-friendly they are. With these insights, we describe, how an ease-to-use emulator is implemented and may run as a Network Emulator as a Service (NEaaS). Therefore, virtualization plays a major role in order to deploy a NEaaS based on Kathará.
After the recent emergence of SARS-CoV-2 infection, unanswered questions remain related to its evolutionary history, path of transmission or divergence and role of recombination. There is emerging evidence on amino acid substitutions occurring in key residues of the receptor-binding domain of the spike glycoprotein in coronavirus isolates from bat and pangolins. In this article, we summarize our current knowledge on the origin of SARS-CoV-2. We also analyze the host ACE2-interacting residues of the receptor-binding domain of spike glycoprotein in SARS-CoV-2 isolates from bats, and compare it to pangolin SARS-CoV-2 isolates collected from Guangdong province (GD Pangolin-CoV) and Guangxi autonomous regions (GX Pangolin-CoV) of South China. Based on our comparative analysis, we support the view that the Guangdong Pangolins are the intermediate hosts that adapted the SARS-CoV-2 and represented a significant evolutionary link in the path of transmission of SARS-CoV-2 virus. We also discuss the role of intermediate hosts in the origin of Omicron.
Measurements of physiological parameters provide an objective, often non-intrusive, and (at least semi-)automatic evaluation and utilization of user behavior. In addition, specific hardware devices of Virtual Reality (VR) often ship with built-in sensors, i.e. eye-tracking and movements sensors. Hence, the combination of physiological measurements and VR applications seems promising. Several approaches have investigated the applicability and benefits of this combination for various fields of applications. However, the range of possible application fields, coupled with potentially useful and beneficial physiological parameters, types of sensor, target variables and factors, and analysis approaches and techniques is manifold. This article provides a systematic overview and an extensive state-of-the-art review of the usage of physiological measurements in VR. We identified 1,119 works that make use of physiological measurements in VR. Within these, we identified 32 approaches that focus on the classification of characteristics of experience, common in VR applications. The first part of this review categorizes the 1,119 works by field of application, i.e. therapy, training, entertainment, and communication and interaction, as well as by the specific target factors and variables measured by the physiological parameters. An additional category summarizes general VR approaches applicable to all specific fields of application since they target typical VR qualities. In the second part of this review, we analyze the target factors and variables regarding the respective methods used for an automatic analysis and, potentially, classification. For example, we highlight which measurement setups have been proven to be sensitive enough to distinguish different levels of arousal, valence, anxiety, stress, or cognitive workload in the virtual realm. This work may prove useful for all researchers wanting to use physiological data in VR and who want to have a good overview of prior approaches taken, their benefits and potential drawbacks.
At the center of the Internet’s protocol stack stands the Internet Protocol (IP) as a common denominator that enables all communication. To make routing efficient, resilient, and scalable, several aspects must be considered. Care must be taken that traffic is well balanced to make efficient use of the existing network resources, both in failure free operation and in failure scenarios.
Finding the optimal routing in a network is an NP-complete problem. Therefore, routing optimization is usually performed using heuristics. This dissertation shows that a routing optimized with one objective function is often not good when looking at other objective functions. It can even be worse than unoptimized routing with respect to that objective function. After looking at failure-free routing and traffic distribution in different failure scenarios, the analysis is extended to include the loop-free alternate (LFA) IP fast reroute mechanism. Different application scenarios of LFAs are examined and a special focus is set on the fact that LFAs usually cannot protect all traffic in a network even against single link failures. Thus, the routing optimization for LFAs is targeted on both link utilization and failure coverage. Finally, the pre-congestion notification mechanism PCN for network admission control and overload protection is analyzed and optimized. Different design options for implementing the protocol are compared, before algorithms are developed for the calculation and optimization of protocol parameters and PCN-based routing.
The second part of the thesis tackles a routing problem that can only be resolved on a global scale. The scalability of the Internet is at risk since a major and intensifying growth of the interdomain routing tables has been observed. Several protocols and architectures are analyzed that can be used to make interdomain routing more scalable. The most promising approach is the locator/identifier (Loc/ID) split architecture which separates routing from host identification. This way, changes in connectivity, mobility of end hosts, or traffic-engineering activities are hidden from the routing in the core of the Internet and the routing tables can be kept much smaller. All of the currently proposed Loc/ID split approaches have their downsides. In particular, the fact that most architectures use the ID for routing outside the Internet’s core is a poor design, which inhibits many of the possible features of a new routing architecture. To better understand the problems and to provide a solution for a scalable routing design that implements a true Loc/ID split, the new GLI-Split protocol is developed in this thesis, which provides separation of global and local routing and uses an ID that is independent from any routing decisions.
Besides GLI-Split, several other new routing architectures implementing Loc/ID split have been proposed for the Internet. Most of them assume that a mapping system is queried for EID-to-RLOC mappings by an intermediate node at the border of an edge network. When the mapping system is queried by an intermediate node, packets are already on their way towards their destination, and therefore, the mapping system must be fast, scalable, secure, resilient, and should be able to relay packets without locators to nodes that can forward them to the correct destination. The dissertation develops a classification for all proposed mapping system architectures and shows their similarities and differences. Finally, the fast two-level mapping system FIRMS is developed. It includes security and resilience features as well as a relay service for initial packets of a flow when intermediate nodes encounter a cache miss for the EID-to-RLOC mapping.
Web caches often use a Time-to-live (TTL) limit to validate data consistency with web servers. We study the impact of TTL constraints on the hit ratio of basic strategies in caches of fixed size. We derive analytical results and confirm their accuracy in comparison to simulations. We propose a score-based caching method with awareness of the current TTL per data for improving the hit ratio close to the upper bound.
Wireless communication is nothing new. The first data transmissions based on electromagnetic waves have been successfully performed at the end of the 19th century. However, it took almost another century until the technology was ripe for mass market. The first mobile communication systems based on the transmission of digital data were introduced in the late 1980s. Within just a couple of years they have caused a revolution in the way people communicate. The number of cellular phones started to outnumber the fixed telephone lines in many countries and is still rising. New technologies in 3G systems, such as UMTS, allow higher data rates and support various kinds of multimedia services. Nevertheless, the end of the road in wireless communication is far from being reached. In the near future, the Internet and cellular phone systems are expected to be integrated to a new form of wireless system. Bandwidth requirements for a rich set of wireless services, e.g.\ video telephony, video streaming, online gaming, will be easily met. The transmission of voice data will just be another IP based service. On the other hand, building such a system is by far not an easy task. The problems in the development of the UMTS system showed the high complexity of wireless systems with support for bandwidth-hungry, IP-based services. But the technological challenges are just one difficulty. Telecommunication systems are planned on a world-wide basis, such that standard bodies, governments, institutions, hardware vendors, and service providers have to find agreements and compromises on a number of different topics. In this work, we provide the reader with a discussion of many of the topics involved in the planning of a Wireless LAN system that is capable of being integrated into the 4th generation mobile networks (4G) that is being discussed nowadays. Therefore, it has to be able to cope with interactive voice and video traffic while still offering high data rates for best effort traffic. Let us assume a scenario where a huge office complex is completely covered with Wireless LAN access points. Different antenna systems are applied in order to reduce the number of access points that are needed on the one hand, while optimizing the coverage on the other. No additional infrastructure is implemented. Our goal is to evaluate whether the Wireless LAN technology is capable of dealing with the various demands of such a scenario. First, each single access point has to be capable of supporting best-effort and Quality of Service (QoS) demanding applications simultaneously. The IT infrastructure in our scenario consists solely of Wireless LAN, such that it has to allow users surfing the Web, while others are involved in voice calls or video conferences. Then, there is the problem of overlapping cells. Users attached to one access point produce interference for others. However, the QoS support has to be maintained, which is not an easy task. Finally, there are nomadic users, which roam from one Wireless LAN cell to another even during a voice call. There are mechanisms in the standard that allow for mobility, but their capabilities for QoS support are yet to be studied. This shows the large number of unresolved issues when it comes to Wireless LAN in the context of 4G networks. In this work we want to tackle some of the problems.
Visual stimuli are frequently used to improve memory, language learning or perception, and understanding of metacognitive processes. However, in virtual reality (VR), there are few systematically and empirically derived databases. This paper proposes the first collection of virtual objects based on empirical evaluation for inter-and transcultural encounters between English- and German-speaking learners. We used explicit and implicit measurement methods to identify cultural associations and the degree of stereotypical perception for each virtual stimuli (n = 293) through two online studies, including native German and English-speaking participants. The analysis resulted in a final well-describable database of 128 objects (called InteractionSuitcase). In future applications, the objects can be used as a great interaction or conversation asset and behavioral measurement tool in social VR applications, especially in the field of foreign language education. For example, encounters can use the objects to describe their culture, or teachers can intuitively assess stereotyped attitudes of the encounters.
This study provides a systematic literature review of research (2001–2020) in the field of teaching and learning a foreign language and intercultural learning using immersive technologies. Based on 2507 sources, 54 articles were selected according to a predefined selection criteria. The review is aimed at providing information about which immersive interventions are being used for foreign language learning and teaching and where potential research gaps exist. The papers were analyzed and coded according to the following categories: (1) investigation form and education level, (2) degree of immersion, and technology used, (3) predictors, and (4) criterions. The review identified key research findings relating the use of immersive technologies for learning and teaching a foreign language and intercultural learning at cognitive, affective, and conative levels. The findings revealed research gaps in the area of teachers as a target group, and virtual reality (VR) as a fully immersive intervention form. Furthermore, the studies reviewed rarely examined behavior, and implicit measurements related to inter- and trans-cultural learning and teaching. Inter- and transcultural learning and teaching especially is an underrepresented investigation subject. Finally, concrete suggestions for future research are given. The systematic review contributes to the challenge of interdisciplinary cooperation between pedagogy, foreign language didactics, and Human-Computer Interaction to achieve innovative teaching-learning formats and a successful digital transformation.
While developing modern applications, it is necessary to ensure an efficient and performant communication between different applications. In current environments, a middleware software is used, which supports the publish/subscribe communication pattern. Using this communication pattern, a publisher sends information encapsulated in messages to the middleware. A subscriber registers its interests at the middleware. The monograph describes three different steps to determine the performance of such a system. In a first step, the message throughput performance of a publish/subscribe in different scenarios is measured using a Java Message Service (JMS) based implementation. In the second step the maximum achievable message throughput is described by adapted models depending on the filter complexity and the replication grade. Using the model, the performance characteristics of a specific system in a given scenario can be determined. These numbers are used for the queuing model described in the third part of the thesis, which supports the dimensioning of a system in realistic scenarios. Additionally, we introduce a method to approximate an M/G/1 system numerically in an efficient way, which can be used for real time analysis to predict the expected performance in a certain scenario. Finally, the analytical model is used to investigate different possibilities to ensure the scalability of the maximum achievable message throughput of the overall system.
CLIP knows image aesthetics
(2022)
Most Image Aesthetic Assessment (IAA) methods use a pretrained ImageNet classification model as a base to fine-tune. We hypothesize that content classification is not an optimal pretraining task for IAA, since the task discourages the extraction of features that are useful for IAA, e.g., composition, lighting, or style. On the other hand, we argue that the Contrastive Language-Image Pretraining (CLIP) model is a better base for IAA models, since it has been trained using natural language supervision. Due to the rich nature of language, CLIP needs to learn a broad range of image features that correlate with sentences describing the image content, composition, environments, and even subjective feelings about the image. While it has been shown that CLIP extracts features useful for content classification tasks, its suitability for tasks that require the extraction of style-based features like IAA has not yet been shown. We test our hypothesis by conducting a three-step study, investigating the usefulness of features extracted by CLIP compared to features obtained from the last layer of a comparable ImageNet classification model. In each step, we get more computationally expensive. First, we engineer natural language prompts that let CLIP assess an image's aesthetic without adjusting any weights in the model. To overcome the challenge that CLIP's prompting only is applicable to classification tasks, we propose a simple but effective strategy to convert multiple prompts to a continuous scalar as required when predicting an image's mean aesthetic score. Second, we train a linear regression on the AVA dataset using image features obtained by CLIP's image encoder. The resulting model outperforms a linear regression trained on features from an ImageNet classification model. It also shows competitive performance with fully fine-tuned networks based on ImageNet, while only training a single layer. Finally, by fine-tuning CLIP's image encoder on the AVA dataset, we show that CLIP only needs a fraction of training epochs to converge, while also performing better than a fine-tuned ImageNet model. Overall, our experiments suggest that CLIP is better suited as a base model for IAA methods than ImageNet pretrained networks.
Radiation therapy today, on account of improvements in treatment procedures over the last 60 years, allows precise treatment of static tumors inside the human body. However, irradiation of moving tumors is still a challenging task as moving tumors often leave the treatment beam and the radiation dose delivered to the tumor reduces simultaneously increasing that on healthy tissue. This research work aims to push the frontiers of radiation therapy in order to enable precise treatment of moving tumors with focus on research and development of a unique real-time system enabling active motion compensation through robotic means to compensate tumor motion. During treatment, patients lie on a treatment couch which is normally used for static position corrections of patient set-up errors prior to radiation treatment. The treatment couch used, called HexaPOD, is a parallel manipulator with six degrees of freedom which can precisely position heavy loads inside a small region. Despite the HexaPOD not initially built with dynamics in mind, it is used in this work for sustained motion compensation by moving patients such that tumors stay precisely located at the center of the treatment beam during the complete course of treatment. In order to realize real-time tumor motion compensation by means of the HexaPOD, several challanges need to be addressed. Real-time aspects are covered by the adoption of a hard real-time operation system in combination with measurement and estimation of latencies of all physical quantities in the compensation system such as tumor or breathing position measurements. Accurate timing information is respected consistently in the whole system and all software-induced latencies are adaptively compensated for. This requires knowledge of future tumor positions from predictors. Several predictors for breathing and tumor motion predictions are proposed and evaluated in terms of a variety of different performance metrics. Extensions to prediction algorithms are introduced fusing both breathing and tumor position information to allow for predictions without the need of an explicit correlation model. Predictions determine the future motion path of the HexaPOD in order to compensate for tumor motion. Several control schemes are developed to enable reference tracking for the HexaPOD. Based on linear and non-linear dynamic modelling of the HexaPOD with system identification methods, a first controller is derived in the form of a model predictive controller. A second controller is proposed based on an assumption of the working principle of the HexaPOD's internal controller. Finally, a third controller is derived as combination of the first and second one. For each of these controllers, comparative results with real hardware experiments and humans in the loop as well as choices of free parameters are presented and discussed. Apart from precise tracking, emphasis is placed on patient comfort which is of crucial importance for acceptance of the system. It is demonstrated that smooth trajectories can be realized by the controllers to guarantee that patients feel comfortable while their tumor motion is compensated at sub-millimeter accuracies. Overall errors of the system are analyzed by relating them to tracking and prediction errors. By exploiting the properties of different predictors, it is shown that the startup time until tracking is reached can be reduced to only a few seconds, even in the case of an initially at-rest HexaPOD and with no initial knowledge of tumor motion. This makes the system especially suitable for the relatively short-fractionated treatment sessions for lung tumors. The tumor motion compensation system has been developed solely based on standard clinical hardware, found in most treatment rooms. With a simple and flexible design, existing treatment can be updated in a cost-efficient way to introduce motion compensation capabilities. Simultaneously, the system does not impose any constraints on state-of-the-art treatment types such as intensity modulated radiotherapy or volumetric modulated arc therapy. Supporting different compensation modes, the system can be applied to any moving tumor whether its motion is predictable (lung tumors) or unpredictable (prostate tumors). By integration of adequate tumor position determination methods, the system can be easily extended to other tumors as well.
Cooperative, connected and automated mobility (CCAM) systems depend on a reliable communication to provide their service and more crucially to ensure the safety of users. One way to ensure the reliability of a data transmission is to use multiple transmission technologies in combination with redundant flows. In this paper, we describe a system requiring multipath communication in the context of CCAM. To this end, we introduce a data plane-based scheduler that uses replication and integration modules to provide redundant and transparent multipath communication. We provide an analytical model for the full replication module of the system and give an overview of how and where the data-plane scheduler components can be realized.
Cooperative, connected and automated mobility (CCAM) systems depend on a reliable communication to provide their service and more crucially to ensure the safety of users. One way to ensure the reliability of a data transmission is to use multiple transmission technologies in combination with redundant flows. In this paper, we describe a system requiring multipath communication in the context of CCAM. To this end, we introduce a data plane-based scheduler that uses replication and integration modules to provide redundant and transparent multipath communication. We provide an analytical model for the full replication module of the system and give an overview of how and where the data-plane scheduler components can be realized.
This work focuses on coordination methods and the control of motion in groups of nonholonomic wheeled mobile robots, in particular of the car-like type. These kind of vehicles are particularly restricted in their mobility. In the main part of this work the two problems of formation motion coordination and of rendezvous in distributed multi-vehicle systems are considered. We introduce several enhancements to an existing motion planning approach for formations of nonholonomic mobile robots. Compared to the original method, the extended approach is able to handle time-varying reference speeds as well as adjustments of the formation's shape during reference trajectory segments with continuously differentiable curvature. Additionally, undesired discontinuities in the speed and steering profiles of the vehicles are avoided. Further, the scenario of snow shoveling on an airfield by utilizing multiple formations of autonomous snowplows is discussed. We propose solutions to the subproblems of motion planning for the formations and tracking control for the individual vehicles. While all situations that might occur have been tested in a simulation environment, we also verified the developed tracking controller in real robot hardware experiments. The task of the rendezvous problem in groups of car-like robots is to drive all vehicles to a common position by means of decentralized control laws. Typically there exists no direct interaction link between all of the vehicles. In this work we present decentralized rendezvous control laws for vehicles with free and with bounded steering. The convergence properties of the approaches are analyzed by utilizing Lyapunov based techniques. Furthermore, they are evaluated within various simulation experiments, while the bounded steering case is also verified within laboratory hardware experiments. Finally we introduce a modification to the bounded steering system that increases the convergence speed at the expense of a higher traveled distance of the vehicles.
Crowdsensing offers a cost-effective way to collect large amounts of environmental sensor data; however, the spatial distribution of crowdsensing sensors can hardly be influenced, as the participants carry the sensors, and, additionally, the quality of the crowdsensed data can vary significantly. Hybrid systems that use mobile users in conjunction with fixed sensors might help to overcome these limitations, as such systems allow assessing the quality of the submitted crowdsensed data and provide sensor values where no crowdsensing data are typically available. In this work, we first used a simulation study to analyze a simple crowdsensing system concerning the detection performance of spatial events to highlight the potential and limitations of a pure crowdsourcing system. The results indicate that even if only a small share of inhabitants participate in crowdsensing, events that have locations correlated with the population density can be easily and quickly detected using such a system. On the contrary, events with uniformly randomly distributed locations are much harder to detect using a simple crowdsensing-based approach. A second evaluation shows that hybrid systems improve the detection probability and time. Finally, we illustrate how to compute the minimum number of fixed sensors for the given detection time thresholds in our exemplary scenario.
Routing is one of the most important issues in any communication network. It defines on which path packets are transmitted from the source of a connection to the destination. It allows to control the distribution of flows between different locations in the network and thereby is a means to influence the load distribution or to reach certain constraints imposed by particular applications. As failures in communication networks appear regularly and cannot be completely avoided, routing is required to be resilient against such outages, i.e., routing still has to be able to forward packets on backup paths even if primary paths are not working any more.
Throughout the years, various routing technologies have been introduced that are very different in their control structure, in their way of working, and in their ability to handle certain failure cases. Each of the different routing approaches opens up their own specific questions regarding configuration, optimization, and inclusion of resilience issues. This monograph investigates, with the example of three particular routing technologies, some concrete issues regarding the analysis and optimization of resilience. It thereby contributes to a better general, technology-independent understanding of these approaches and of their diverse potential for the use in future network architectures.
The first considered routing type, is decentralized intra-domain routing based on administrative IP link costs and the shortest path principle. Typical examples are common today's intra-domain routing protocols OSPF and IS-IS. This type of routing includes automatic restoration abilities in case of failures what makes it in general very robust even in the case of severe network outages including several failed components. Furthermore, special IP-Fast Reroute mechanisms allow for a faster reaction on outages. For routing based on link costs, traffic engineering, e.g. the optimization of the maximum relative link load in the network, can be done indirectly by changing the administrative link costs to adequate values.
The second considered routing type, MPLS-based routing, is based on the a priori configuration of primary and backup paths, so-called Label Switched Paths. The routing layout of MPLS paths offers more freedom compared to IP-based routing as it is not restricted by any shortest path constraints but any paths can be setup. However, this in general involves a higher configuration effort.
Finally, in the third considered routing type, typically centralized routing using a Software Defined Networking (SDN) architecture, simple switches only forward packets according to routing decisions made by centralized controller units. SDN-based routing layouts offer the same freedom as for explicit paths configured using MPLS. In case of a failure, new rules can be setup by the controllers to continue the routing in the reduced topology. However, new resilience issues arise caused by the centralized architecture. If controllers are not reachable anymore, the forwarding rules in the single nodes cannot be adapted anymore. This might render a rerouting in case of connection problems in severe failure scenarios infeasible.
Within the healthcare environment, mobile health (mHealth) applications (apps) are becoming more and more important. The number of new mHealth apps has risen steadily in the last years. Especially the COVID-19 pandemic has led to an enormous amount of app releases. In most countries, mHealth applications have to be compliant with several regulatory aspects to be declared a “medical app”. However, the latest applicable medical device regulation (MDR) does not provide more details on the requirements for mHealth applications. When developing a medical app, it is essential that all contributors in an interdisciplinary team — especially software engineers — are aware of the specific regulatory requirements beforehand. The development process, however, should not be stalled due to integration of the MDR. Therefore, a developing framework that includes these aspects is required to facilitate a reliable and quick development process. The paper at hand introduces the creation of such a framework on the basis of the Corona Health and Corona Check apps. The relevant regulatory guidelines are listed and summarized as a guidance for medical app developments during the pandemic and beyond. In particular, the important stages and challenges faced that emerged during the entire development process are highlighted.
This thesis deals with the management and analysis of source code, which is represented in XML. Using the elementary methods of the XML repository, the XML source code representation is accessed, changed, updated, and saved. We reason about the source code, refactor source code and we visualize dependency graphs for call analysis. The visualized dependencies between files, modules, or packages are used to structure the source code in order to get a system, which is easily to comprehend, to modify and to complete. Sophisticated methods have been developed to slice the source code in order to obtain a working package of a large system, containing only a specific functionality. The basic methods, on which the visualizations and analyses are built on can be changed like changing a plug-in. The visualization methods can be reused in order to handle arbitrary source code representations, e.g., JAML, PHPML, PROLOGML. Dependencies of other context can be visualized, too, e.g., ER diagrams, or website references. The tool SCAV supports source code visualization and analyzing methods.
Scalability is often mentioned in literature, but a stringent definition is missing. In particular, there is no general scalability assessment which clearly indicates whether a system scales or not or whether a system scales better than another. The key contribution of this article is the definition of a scalability index (SI) which quantifies if a system scales in comparison to another system, a hypothetical system, e.g., linear system, or the theoretically optimal system. The suggested SI generalizes different metrics from literature, which are specialized cases of our SI. The primary target of our scalability framework is, however, benchmarking of two systems, which does not require any reference system. The SI is demonstrated and evaluated for different use cases, that are (1) the performance of an IoT load balancer depending on the system load, (2) the availability of a communication system depending on the size and structure of the network, (3) scalability comparison of different location selection mechanisms in fog computing with respect to delays and energy consumption; (4) comparison of time-sensitive networking (TSN) mechanisms in terms of efficiency and utilization. Finally, we discuss how to use and how not to use the SI and give recommendations and guidelines in practice. To the best of our knowledge, this is the first work which provides a general SI for the comparison and benchmarking of systems, which is the primary target of our scalability analysis.
3D point clouds are a de facto standard for 3D documentation and modelling. The advances in laser scanning technology broadens the usability and access to 3D measurement systems. 3D point clouds are used in many disciplines such as robotics, 3D modelling, archeology and surveying. Scanners are able to acquire up to a million of points per second to represent the environment with a dense point cloud. This represents the captured environment with a very high degree of detail. The combination of laser scanning technology with photography adds color information to the point clouds. Thus the environment is represented more realistically. Full 3D models of environments, without any occlusion, require multiple scans. Merging point clouds is a challenging process. This thesis presents methods for point cloud registration based on the panorama images generated from the scans. Image representation of point clouds introduces 2D image processing methods to 3D point clouds. Several projection methods for the generation of panorama maps of point clouds are presented in this thesis. Additionally, methods for point cloud reduction and compression based on the panorama maps are proposed. Due to the large amounts of data generated from the 3D measurement systems these methods are necessary to improve the point cloud processing, transmission and archiving. This thesis introduces point cloud processing methods as a novel framework for the digitisation of archeological excavations. The framework replaces the conventional documentation methods for excavation sites. It employs point clouds for the generation of the digital documentation of an excavation with the help of an archeologist on-site. The 3D point cloud is used not only for data representation but also for analysis and knowledge generation. Finally, this thesis presents an autonomous indoor mobile mapping system. The mapping system focuses on the sensor placement planning method. Capturing a complete environment requires several scans. The sensor placement planning method solves for the minimum required scans to digitise large environments. Combining this method with a navigation system on a mobile robot platform enables it to acquire data fully autonomously. This thesis introduces a novel hole detection method for point clouds to detect obscured parts of a captured environment. The sensor placement planning method selects the next scan position with the most coverage of the obscured environment. This reduces the required number of scans. The navigation system on the robot platform consist of path planning, path following and obstacle avoidance. This guarantees the safe navigation of the mobile robot platform between the scan positions. The sensor placement planning method is designed as a stand alone process that could be used with a mobile robot platform for autonomous mapping of an environment or as an assistant tool for the surveyor on scanning projects.
In future telecommunication systems, we observe an increasing diversity of access networks. The separation of transport services and applications or services leads to multi-network services, i.e., a future service has to work transparently to the underlying network infrastructure. Multi-network services with edge-based intelligence, like P2P file sharing or the Skype VoIP service, impose new traffic control paradigms on the future Internet. Such services adapt the amount of consumed bandwidth to reach different goals. A selfish behavior tries to keep the QoE of a single user above a certain level. Skype, for instance, repeats voice samples depending on the perceived end-to-end loss. From the viewpoint of a single user, the replication of voice data overcomes the degradation caused by packet loss and enables to maintain a certain QoE. The cost for this achievement is a higher amount of consumed bandwidth. However, if the packet loss is caused by congestion in the network, this additionally required bandwidth even worsens the network situation. Altruistic behavior, on the other side, would reduce the bandwidth consumption in such a way that the pressure on the network is released and thus the overall network performance is improved. In this monograph, we analyzed the impact of the overlay, P2P, and QoE paradigms in future Internet applications and the interactions from the observing user behavior. The shift of intelligence toward the edge is accompanied by a change in the emerging user behavior and traffic profile, as well as a change from multi-service networks to multi-networks services. In addition, edge-based intelligence may lead to a higher dynamics in the network topology, since the applications are often controlled by an overlay network, which can rapidly change in size and structure as new nodes can leave or join the overlay network in an entirely distributed manner. As a result, we found that the performance evaluation of such services provides new challenges, since novel key performance factors have to be first identified, like pollution of P2P systems, and appropriate models of the emerging user behavior are required, e.g. taking into account user impatience. As common denominator of the presented studies in this work, we focus on a user-centric view when evaluating the performance of future Internet applications. For a subscriber of a certain application or service, the perceived quality expressed as QoE will be the major criterion of the user's satisfaction with the network and service providers. We selected three different case studies and characterized the application's performance from the end user's point of view. Those are (1) cooperation in mobile P2P file sharing networks, (2) modeling of online TV recording services, and (3) QoE of edge-based VoIP applications. The user-centric approach facilitates the development of new mechanisms to overcome problems arising from the changing user behavior. An example is the proposed CycPriM cooperation strategy, which copes with selfish user behavior in mobile P2P file sharing system. An adequate mechanism has also been shown to be efficient in a heterogeneous B3G network with mobile users conducting vertical handovers between different wireless access technologies. The consideration of the user behavior and the user perceived quality guides to an appropriate modeling of future Internet applications. In the case of the online TV recording service, this enables the comparison between different technical realizations of the system, e.g. using server clusters or P2P technology, to properly dimension the installed network elements and to assess the costs for service providers. Technologies like P2P help to overcome phenomena like flash crowds and improve scalability compared to server clusters, which may get overloaded in such situations. Nevertheless, P2P technology invokes additional challenges and different user behavior to that seen in traditional client/server systems. Beside the willingness to share files and the churn of users, peers may be malicious and offer fake contents to disturb the data dissemination. Finally, the understanding and the quantification of QoE with respect to QoS degradations permits designing sophisticated edge-based applications. To this end, we identified and formulated the IQX hypothesis as an exponential interdependency between QoE and QoS parameters, which we validated for different examples. The appropriate modeling of the emerging user behavior taking into account the user's perceived quality and its interactions with the overlay and P2P paradigm will finally help to design future Internet applications.
For formative evaluations of user experience (UX) a variety of methods have been developed over the years. However, most techniques require the users to interact with the study as a secondary task. This active involvement in the evaluation is not inclusive of all users and potentially biases the experience currently being studied. Yet there is a lack of methods for situations in which the user has no spare cognitive resources. This condition occurs when 1) users' cognitive abilities are impaired (e.g., people with dementia) or 2) users are confronted with very demanding tasks (e.g., air traffic controllers). In this work we focus on emotions as a key component of UX and propose the new structured observation method Proxemo for formative UX evaluations. Proxemo allows qualified observers to document users' emotions by proxy in real time and then directly link them to triggers. Technically this is achieved by synchronising the timestamps of emotions documented by observers with a video recording of the interaction.
In order to facilitate the documentation of observed emotions in highly diverse contexts we conceptualise and implement two separate versions of a documentation aid named Proxemo App. For formative UX evaluations of technology-supported reminiscence sessions with people with dementia, we create a smartwatch app to discreetly document emotions from the categories anger, general alertness, pleasure, wistfulness and pride. For formative UX evaluations of prototypical user interfaces with air traffic controllers we create a smartphone app to efficiently document emotions from the categories anger, boredom, surprise, stress and pride. Descriptive case studies in both application domains indicate the feasibility and utility of the method Proxemo and the appropriateness of the respectively adapted design of the Proxemo App.
The third part of this work is a series of meta-evaluation studies to determine quality criteria of Proxemo. We evaluate Proxemo regarding its reliability, validity, thoroughness and effectiveness, and compare Proxemo's efficiency and the observers' experience to documentation with pen and paper. Proxemo is reliable, as well as more efficient, thorough and effective than handwritten notes and provides a better UX to observers. Proxemo compares well with existing methods where benchmarks are available.
With Proxemo we contribute a validated structured observation method that has shown to meet requirements formative UX evaluations in the extreme contexts of users with cognitive impairments or high task demands. Proxemo is agnostic regarding researchers' theoretical approaches and unites reductionist and holistic perspectives within one method.
Future work should explore the applicability of Proxemo for further domains and extend the list of audited quality criteria to include, for instance, downstream utility. With respect to basic research we strive to better understand the sources leading observers to empathic judgments and propose reminisce and older adults as model environment for investigating mixed emotions.
Two studies are reported that investigate how readily accessible and applicable ten force-dynamic categories are to novices in describing short episodes of human-technology interaction (Study 1) and that establish a measure of inter-coder reliability when re-classifying these episodes into force-dynamic categories (Study 2). The results of the first study show that people can easily and confidently relate their experiences with technology to the definitions of force-dynamic events (e.g. “The driver released the handbrake” as an example of restraint removal). The results of the second study show moderate agreement between four expert coders across all ten force-dynamic categories (Cohen’s kappa = .59) when re-classifying these episodes. Agreement values for single force-dynamic categories ranged between ‘fair’ and ‘almost perfect’, i.e. between kappa = .30 and .95. Agreement with the originally intended classifications of study 1 was higher than the pure inter-coder reliabilities. Single coders achieved an average kappa of .71, indicating substantial agreement. Using more than one coder increased kappas to almost perfect: up to .87 for four coders. A qualitative analysis of the predicted versus the observed number of category confusions revealed that about half of the category disagreement could be predicted from strong overlaps in the definitions of force-dynamic categories. From the quantitative and qualitative results, guidelines are derived to aid the better training of coders in order to increase inter-coder reliability.
Future mobile communication networks, such as 5G and beyond, can benefit from Virtualized Network Functions (VNFs) when deployed on cloud infrastructures to achieve elasticity and scalability. However, new challenges arise as to managing states of Network Functions (NFs). Especially control plane VNFs, which are mainly found in cellular core networks like the 5G Core (5GC), received little attention since the shift towards virtualizing NFs. Most existing solutions for these core networks are often complex, intrusive, and are seldom compliant with the standard. With the emergence of 5G campus networks, UEs will be mainly machine-type devices. These devices communicate more deterministically, bringing new opportunities for elaborated state management. This work presents an emulation environment to perform rigorous measurements on state access patterns. The emulation comes with a fully parameterized Markov model for the UE to examine a wide variety of different devices. These measurements can then be used as a solid base for designing an efficient, simple, and standard conform state management solution that brings us further towards stateless core networks.
With the introduction of OpenFlow by the Stanford University in 2008, a process began in the area of network research, which questions the predominant approach of fully distributed network control. OpenFlow is a communication protocol that allows the externalization of the network control plane from the network devices, such as a router, and to realize it as a logically-centralized entity in software. For this concept, the term "Software Defined Networking" (SDN) was coined during scientific discourse.
For the network operators, this concept has several advantages. The two most important can be summarized under the points cost savings and flexibility. Firstly, it is possible through the uniform interface for network hardware ("Southbound API"), as implemented by OpenFlow, to combine devices and software from different manufacturers, which increases the innovation and price pressure on them. Secondly, the realization of the network control plane as a freely programmable software with open interfaces ("Northbound API") provides the opportunity to adapt it to the individual circumstances of the operator's network and to exchange information with the applications it serves. This allows the network to be more flexible and to react more quickly to changing circumstances as well as transport the traffic more effectively and tailored to the user’s "Quality of Experience" (QoE).
The approach of a separate network control layer for packet-based networks is not new and has already been proposed several times in the past. Therefore, the SDN approach has raised many questions about its feasibility in terms of efficiency and applicability. These questions are caused to some extent by the fact that there is no generally accepted definition of the SDN concept to date. It is therefore a part of this thesis to derive such a definition. In addition, several of the open issues are investigated. This Investigations follow the three aspects: Performance Evaluation of Software Defined Networking, applications on the SDN control layer, and the usability of SDN Northbound-API for creation application-awareness in network operation.
Performance Evaluation of Software Defined Networking: The question of the efficiency of an SDN-based system was from the beginning one of the most important. In this thesis, experimental measurements of the performance of OpenFlow-enabled switch hardware and control software were conducted for the purpose of answering this question. The results of these measurements were used as input parameters for establishing an analytical model of the reactive SDN approach. Through the model it could be determined that the performance of the software control layer, often called "Controller", is crucial for the overall performance of the system, but that the approach is generally viable. Based on this finding a software for analyzing the performance of SDN controllers was developed. This software allows the emulation of the forwarding layer of an SDN network towards the control software and can thus determine its performance in different situations and configurations. The measurements with this software showed that there are quite significant differences in the behavior of different control software implementations. Among other things it has been shown that some show different characteristics for various switches, in particular in terms of message processing speed. Under certain circumstances this can lead to network failures.
Applications on the SDN control layer: The core piece of software defined networking are the intelligent network applications that operate on the control layer. However, their development is still in its infancy and little is known about the technical possibilities and their limitations. Therefore, the relationship between an SDN-based and classical implementation of a network function is investigated in this thesis. This function is the monitoring of network links and the traffic they carry. A typical approach for this task has been built based on Wiretapping and specialized measurement hardware and compared with an implementation based on OpenFlow switches and a special SDN control application. The results of the comparison show that the SDN version can compete in terms of measurement accuracy for bandwidth and delay estimation with the traditional measurement set-up. However, a compromise has to be found for measurements below the millisecond range.
Another question regarding the SDN control applications is whether and how well they can solve existing problems in networks. Two programs have been developed based on SDN in this thesis to solve two typical network issues. Firstly, the tool "IPOM", which enables considerably more flexibility in the study of effects of network structures for a researcher, who is confined to a fixed physical test network topology.
The second software provides an interface between the Cloud Orchestration Software "OpenNebula" and an OpenFlow controller. The purpose of this software was to investigate experimentally whether a pre-notification of the network of an impending relocation of a virtual service in a data center is sufficient to ensure the continuous operation of that service. This was demonstrated on the example of a video service.
Usability of the SDN Northbound API for creating application-awareness in network operation: Currently, the fact that the network and the applications that run on it are developed and operated separately leads to problems in network operation. SDN offers with the Northbound-API an open interface that enables the exchange between information of both worlds during operation. One aim of this thesis was to investigate whether this interface can be exploited so that the QoE experienced by the user can be maintained on high level. For this purpose, the QoE influence factors were determined on a challenging application by means of a subjective survey study. The application is cloud gaming, in which the calculation of video game environments takes place in the cloud and is transported via video over the network to the user. It was shown that apart from the most important factor influencing QoS, i.e., packet loss on the downlink, also the type of game type and its speed play a role. This demonstrates that in addition to QoS the application state is important and should be communicated to the network. Since an implementation of such a state conscious SDN for the example of Cloud Gaming was not possible due to its proprietary implementation, in this thesis the application “YouTube video streaming” was chosen as an alternative. For this application, status information is retrievable via the "Yomo" tool and can be used for network control. It was shown that an SDN-based implementation of an application-aware network has distinct advantages over traditional network management methods and the user quality can be obtained in spite of disturbances.
In the present day, unmanned aerial vehicles become seemingly more popular every year, but, without regulation of the increasing number of these vehicles, the air space could become chaotic and uncontrollable. In this work, a framework is proposed to combine self-aware computing with multirotor formations to address this problem. The self-awareness is envisioned to improve the dynamic behavior of multirotors. The formation scheme that is implemented is called platooning, which arranges vehicles in a string behind the lead vehicle and is proposed to bring order into chaotic air space. Since multirotors define a general category of unmanned aerial vehicles, the focus of this thesis are quadcopters, platforms with four rotors. A modification for the LRA-M self-awareness loop is proposed and named Platooning Awareness. The implemented framework is able to offer two flight modes that enable waypoint following and the self-awareness module to find a path through scenarios, where obstacles are present on the way, onto a goal position. The evaluation of this work shows that the proposed framework is able to use self-awareness to learn about its environment, avoid obstacles, and can successfully move a platoon of drones through multiple scenarios.
Automatic image reconstruction is critical to cope with steadily increasing data from advanced microscopy. We describe here the Fiji macro 3D ART VeSElecT which we developed to study synaptic vesicles in electron tomograms. We apply this tool to quantify vesicle properties (i) in embryonic Danio rerio 4 and 8 days past fertilization (dpf) and (ii) to compare Caenorhabditis elegans N2 neuromuscular junctions (NMJ) wild-type and its septin mutant (unc-59(e261)). We demonstrate development-specific and mutant-specific changes in synaptic vesicle pools in both models. We confirm the functionality of our macro by applying our 3D ART VeSElecT on zebrafish NMJ showing smaller vesicles in 8 dpf embryos then 4 dpf, which was validated by manual reconstruction of the vesicle pool. Furthermore, we analyze the impact of C. elegans septin mutant unc-59(e261) on vesicle pool formation and vesicle size. Automated vesicle registration and characterization was implemented in Fiji as two macros (registration and measurement). This flexible arrangement allows in particular reducing false positives by an optional manual revision step. Preprocessing and contrast enhancement work on image-stacks of 1nm/pixel in x and y direction. Semi-automated cell selection was integrated. 3D ART VeSElecT removes interfering components, detects vesicles by 3D segmentation and calculates vesicle volume and diameter (spherical approximation, inner/outer diameter). Results are collected in color using the RoiManager plugin including the possibility of manual removal of non-matching confounder vesicles. Detailed evaluation considered performance (detected vesicles) and specificity (true vesicles) as well as precision and recall. We furthermore show gain in segmentation and morphological filtering compared to learning based methods and a large time gain compared to manual segmentation. 3D ART VeSElecT shows small error rates and its speed gain can be up to 68 times faster in comparison to manual annotation. Both automatic and semi-automatic modes are explained including a tutorial.
Ambalytics: a scalable and distributed system architecture concept for bibliometric network analyses
(2021)
A deep understanding about a field of research is valuable for academic researchers. In addition to technical knowledge, this includes knowledge about subareas, open research questions, and social communities (networks) of individuals and organizations within a given field. With bibliometric analyses, researchers can acquire quantitatively valuable knowledge about a research area by using bibliographic information on academic publications provided by bibliographic data providers. Bibliometric analyses include the calculation of bibliometric networks to describe affiliations or similarities of bibliometric entities (e.g., authors) and group them into clusters representing subareas or communities. Calculating and visualizing bibliometric networks is a nontrivial and time-consuming data science task that requires highly skilled individuals. In addition to domain knowledge, researchers must often provide statistical knowledge and programming skills or use software tools having limited functionality and usability. In this paper, we present the ambalytics bibliometric platform, which reduces the complexity of bibliometric network analysis and the visualization of results. It accompanies users through the process of bibliometric analysis and eliminates the need for individuals to have programming skills and statistical knowledge, while preserving advanced functionality, such as algorithm parameterization, for experts. As a proof-of-concept, and as an example of bibliometric analyses outcomes, the calculation of research fronts networks based on a hybrid similarity approach is shown. Being designed to scale, ambalytics makes use of distributed systems concepts and technologies. It is based on the microservice architecture concept and uses the Kubernetes framework for orchestration. This paper presents the initial building block of a comprehensive bibliometric analysis platform called ambalytics, which aims at a high usability for users as well as scalability.
For machine manufacturing companies, besides the production of high quality and reliable machines, requirements have emerged to maintain machine-related aspects through digital services. The development of such services in the field of the Industrial Internet of Things (IIoT) is dealing with solutions such as effective condition monitoring and predictive maintenance. However, appropriate data sources are needed on which digital services can be technically based. As many powerful and cheap sensors have been introduced over the last years, their integration into complex machines is promising for developing digital services for various scenarios. It is apparent that for components handling recorded data of these sensors they must usually deal with large amounts of data. In particular, the labeling of raw sensor data must be furthered by a technical solution. To deal with these data handling challenges in a generic way, a sensor processing pipeline (SPP) was developed, which provides effective methods to capture, process, store, and visualize raw sensor data based on a processing chain. Based on the example of a machine manufacturing company, the SPP approach is presented in this work. For the company involved, the approach has revealed promising results.
Nowadays, robotics plays an important role in increasing fields of application. There exist many environments or situations where mobile robots instead of human beings are used, since the tasks are too hazardous, uncomfortable, repetitive, or costly for humans to perform. The autonomy and the mobility of the robot are often essential for a good solution of these problems. Thus, such a robot should at least be able to answer the question "Where am I?". This thesis investigates the problem of self-localizing a robot in an indoor environment using range measurements. That is, a robot equipped with a range sensor wakes up inside a building and has to determine its position using only its sensor data and a map of its environment. We examine this problem from an idealizing point of view (reducing it into a pure geometric one) and further investigate a method of Guibas, Motwani, and Raghavan from the field of computational geometry to solving it. Here, so-called visibility skeletons, which can be seen as coarsened representations of visibility polygons, play a decisive role. In the major part of this thesis we analyze the structures and the occurring complexities in the framework of this scheme. It turns out that the main source of complication are so-called overlapping embeddings of skeletons into the map polygon, for which we derive some restrictive visibility constraints. Based on these results we are able to improve one of the occurring complexity bounds in the sense that we can formulate it with respect to the number of reflex vertices instead of the total number of map vertices. This also affects the worst-case bound on the preprocessing complexity of the method. The second part of this thesis compares the previous idealizing assumptions with the properties of real-world environments and discusses the occurring problems. In order to circumvent these problems, we use the concept of distance functions, which model the resemblance between the sensor data and the map, and appropriately adapt the above method to the needs of realistic scenarios. In particular, we introduce a distance function, namely the polar coordinate metric, which seems to be well suited to the localization problem. Finally, we present the RoLoPro software where most of the discussed algorithms are implemented (including the polar coordinate metric).
Background: Boolean networks capture switching behavior of many naturally occurring regulatory networks. For semi-quantitative modeling, interpolation between ON and OFF states is necessary. The high degree polynomial interpolation of Boolean genetic regulatory networks (GRNs) in cellular processes such as apoptosis or proliferation allows for the modeling of a wider range of node interactions than continuous activator-inhibitor models, but suffers from scaling problems for networks which contain nodes with more than ~10 inputs. Many GRNs from literature or new gene expression experiments exceed those limitations and a new approach was developed.
Results: (i) As a part of our new GRN simulation framework Jimena we introduce and setup Boolean-tree-based data structures; (ii) corresponding algorithms greatly expedite the calculation of the polynomial interpolation in almost all cases, thereby expanding the range of networks which can be simulated by this model in reasonable time. (iii) Stable states for discrete models are efficiently counted and identified using binary decision diagrams. As application example, we show how system states can now be sampled efficiently in small up to large scale hormone disease networks (Arabidopsis thaliana development and immunity, pathogen Pseudomonas syringae and modulation by cytokinins and plant hormones).
Conclusions: Jimena simulates currently available GRNs about 10-100 times faster than the previous implementation of the polynomial interpolation model and even greater gains are achieved for large scale-free networks. This speed-up also facilitates a much more thorough sampling of continuous state spaces which may lead to the identification of new stable states. Mutants of large networks can be constructed and analyzed very quickly enabling new insights into network robustness and behavior.
An important but very time consuming part of the research process is literature review. An already large and nevertheless growing ground set of publications as well as a steadily increasing publication rate continue to worsen the situation. Consequently, automating this task as far as possible is desirable. Experimental results of systems are key-insights of high importance during literature review and usually represented in form of tables. Our pipeline KIETA exploits these tables to contribute to the endeavor of automation by extracting them and their contained knowledge from scientific publications. The pipeline is split into multiple steps to guarantee modularity as well as analyzability, and agnosticim regarding the specific scientific domain up until the knowledge extraction step, which is based upon an ontology. Additionally, a dataset of corresponding articles has been manually annotated with information regarding table and knowledge extraction. Experiments show promising results that signal the possibility of an automated system, while also indicating limits of extracting knowledge from tables without any context.
This article introduces the Off-The-Shelf Stylus (OTSS), a framework for 2D interaction (in 3D) as well as for handwriting and sketching with digital pen, ink, and paper on physically aligned virtual surfaces in Virtual, Augmented, and Mixed Reality (VR, AR, MR: XR for short). OTSS supports self-made XR styluses based on consumer-grade six-degrees-of-freedom XR controllers and commercially available styluses. The framework provides separate modules for three basic but vital features: 1) The stylus module provides stylus construction and calibration features. 2) The surface module provides surface calibration and visual feedback features for virtual-physical 2D surface alignment using our so-called 3ViSuAl procedure, and surface interaction features. 3) The evaluation suite provides a comprehensive test bed combining technical measurements for precision, accuracy, and latency with extensive usability evaluations including handwriting and sketching tasks based on established visuomotor, graphomotor, and handwriting research. The framework’s development is accompanied by an extensive open source reference implementation targeting the Unity game engine using an Oculus Rift S headset and Oculus Touch controllers. The development compares three low-cost and low-tech options to equip controllers with a tip and includes a web browser-based surface providing support for interacting, handwriting, and sketching. The evaluation of the reference implementation based on the OTSS framework identified an average stylus precision of 0.98 mm (SD = 0.54 mm) and an average surface accuracy of 0.60 mm (SD = 0.32 mm) in a seated VR environment. The time for displaying the stylus movement as digital ink on the web browser surface in VR was 79.40 ms on average (SD = 23.26 ms), including the physical controller’s motion-to-photon latency visualized by its virtual representation (M = 42.57 ms, SD = 15.70 ms). The usability evaluation (N = 10) revealed a low task load, high usability, and high user experience. Participants successfully reproduced given shapes and created legible handwriting, indicating that the OTSS and it’s reference implementation is ready for everyday use. We provide source code access to our implementation, including stylus and surface calibration and surface interaction features, making it easy to reuse, extend, adapt and/or replicate previous results (https://go.uniwue.de/hci-otss).
Graphs are a frequently used tool to model relationships among entities. A graph is a binary relation between objects, that is, it consists of a set of objects (vertices) and a set of pairs of objects (edges).
Networks are common examples of modeling data as a graph. For example, relationships between persons in a social network, or network links between computers in a telecommunication network can be represented by a graph.
The clearest way to illustrate the modeled data is to visualize the graphs. The field of Graph Drawing deals with the problem of finding algorithms to automatically generate graph visualizations. The task is to find a "good" drawing, which can be measured by different criteria such as number of crossings between edges or the used area. In this thesis, we study Angular Schematization in Graph Drawing. By this, we mean drawings
with large angles (for example, between the edges at common vertices or at crossing points).
The thesis consists of three parts. First, we deal with the placement of boxes. Boxes are axis-parallel rectangles that can, for example, contain text.
They can be placed on a map to label important sites, or can be used to describe semantic relationships between words in a word network. In the second part of the thesis, we consider graph drawings visually guide the
viewer. These drawings generally induce large angles between edges that meet at a vertex. Furthermore, the edges are drawn crossing-free and in a way that
makes them easy to follow for the human eye. The third and final part is devoted to crossings with large angles. In drawings with crossings, it is important to have large angles between edges at their crossing point, preferably right angles.
The Internet of Things (IoT) enables a variety of smart applications, including smart home, smart manufacturing, and smart city. By enhancing Business Process Management Systems with IoT capabilities, the execution and monitoring of business processes can be significantly improved. Providing a holistic support for modeling, executing and monitoring IoT-driven processes, however, constitutes a challenge. Existing process modeling and process execution languages, such as BPMN 2.0, are unable to fully meet the IoT characteristics (e.g., asynchronicity and parallelism) of IoT-driven processes. In this article, we present BPMNE4IoT—A holistic framework for modeling, executing and monitoring IoT-driven processes. We introduce various artifacts and events based on the BPMN 2.0 metamodel that allow realizing the desired IoT awareness of business processes. The framework is evaluated along two real-world scenarios from two different domains. Moreover, we present a user study for comparing BPMNE4IoT and BPMN 2.0. In particular, this study has confirmed that the BPMNE4IoT framework facilitates the support of IoT-driven processes.
The focus of this work lies on the communication issues of Medium Access Control (MAC) and routing protocols in the context of WSNs. The communication challenges in these networks mainly result from high node density, low bandwidth, low energy constraints and the hardware limitations in terms of memory, computational power and sensing capabilities of low-power transceivers. For this reason, the structure of WSNs is always kept as simple as possible to minimize the impact of communication issues. Thus, the majority of WSNs apply a simple one hop star topology since multi-hop communication has high demands on the routing protocol since it increases the bandwidth requirements of the network. Moreover, medium access becomes a challenging problem due to the fact that low-power transceivers are very limited in their sensing capabilities. The first contribution is represented by the Backoff Preamble-based MAC Protocol with Sequential Contention Resolution (BPS-MAC) which is designed to overcome the limitations of low-power transceivers. Two communication issues, namely the Clear Channel Assessment (CCA) delay and the turnaround time, are directly addressed by the protocol. The CCA delay represents the period of time which is required by the transceiver to detect a busy radio channel while the turnaround time specifies the period of time which is required to switch between receive and transmit mode. Standard Carrier Sense Multiple Access (CSMA) protocols do not achieve high performance in terms of packet loss if the traffic is highly correlated due to the fact that the transceiver is not able to sense the medium during the switching phase. Therefore, a node may start to transmit data while another node is already transmitting since it has sensed an idle medium right before it started to switch its transceiver from receive to transmit mode. The BPS-MAC protocol uses a new sequential preamble-based medium access strategy which can be adapted to the hardware capabilities of the transceivers. The protocol achieves a very low packet loss rate even in wireless networks with high node density and event-driven traffic without the need of synchronization. This makes the protocol attractive to applications such as structural health monitoring, where event suppression is not an option. Moreover, acknowledgments or complex retransmission strategies become almost unnecessary since the sequential preamble-based contention resolution mechanism minimizes the collision probability. However, packets can still be lost as a consequence of interference or other issues which affect signal propagation. The second contribution consists of a new routing protocol which is able to quickly detect topology changes without generating a large amount of overhead. The key characteristics of the Statistic-Based Routing (SBR) protocol are high end-to-end reliability (in fixed and mobile networks), load balancing capabilities, a smooth continuous routing metric, quick adaptation to changing network conditions, low processing and memory requirements, low overhead, support of unidirectional links and simplicity. The protocol can establish routes in a hybrid or a proactive mode and uses an adaptive continuous routing metric which makes it very flexible in terms of scalability while maintaining stable routes. The hybrid mode is optimized for low-power WSNs since routes are only established on demand. The difference of the hybrid mode to reactive routing strategies is that routing messages are periodically transmitted to maintain already established routes. However, the protocol stops the transmission of routing messages if no data packets are transmitted for a certain time period in order to minimize the routing overhead and the energy consumption. The proactive mode is designed for high data rate networks which have less energy constraints. In this mode, the protocol periodically transmits routing messages to establish routes in a proactive way even in the absence of data traffic. Thus, nodes in the network can immediately transmit data since the route to the destination is already established in advance. In addition, a new delay-based routing message forwarding strategy is introduced. The forwarding strategy is part of SBR but can also be applied to many routing protocols in order to modify the established topology. The strategy can be used, e.g. in mobile networks, to decrease the packet loss by deferring routing messages with respect to the neighbor change rate. Thus, nodes with a stable neighborhood forward messages faster than nodes within a fast changing neighborhood. As a result, routes are established through nodes with correlated movement which results in fewer topology changes due to higher link durations.
Today’s Internet architecture was not designed from scratch but was driven by new services that emerged during its development. Hence, it is often described as patchwork where additional patches are applied in case new services require modifications to the existing architecture. This process however is rather slow and hinders the development of innovative network services with certain architecture or network requirements. Currently discussed technologies like Software-Defined Networking (SDN) or Network Virtualization (NV) are seen as key enabling technologies to overcome this rigid best effort legacy of the Internet. Both technologies offer the possibility to create virtual networks that accommodate the specific needs of certain services. These logical networks are operated on top of a physical substrate and facilitate flexible network resource allocation as physical resources can be added and removed depending on the current network and load situation. In addition, the clear separation and isolation of networks foster the development of application-aware networks that fulfill the special requirements of emerging applications. A prominent use case that benefits from these extended capabilities of the network is denoted with service component mobility. Services hosted on Virtual Machines (VMs) follow their consuming mobile endpoints, so that access latency as well as consumed network resources are reduced. Especially for applications like video streaming, which consume a large fraction of the available resources, is this an important means to relieve the resource constraints and eventually provide better service quality. Service and endpoint mobility both allow an adaptation of the used paths between an offered service, i.e., video streaming and the consuming users in case the service quality drops due to network problems. To make evidence-based adaptations in case of quality drops, a scalable monitoring component is required that is able to monitor the service quality for video streaming applications with reliable accuracy. This monograph details challenges that arise when deploying a certain service, i.e., video streaming, in a future virtualized network architecture and discusses possible solutions. In particular, this work evaluates the performance of mechanisms enabling service mobility and presents an optimized architecture for service mobility. Concerning endpoint mobility, improvements are developed that reduce the latency between endpoints and consumed services and ensure connectivity regardless of the used mobile access network. In the last part, a network-based video quality monitoring solution is developed and its accuracy is evaluated.
A bipartite graph G=(U,V,E) is convex if the vertices in V can be linearly ordered such that for each vertex u∈U, the neighbors of u are consecutive in the ordering of V. An induced matching H of G is a matching for which no edge of E connects endpoints of two different edges of H. We show that in a convex bipartite graph with n vertices and m weighted edges, an induced matching of maximum total weight can be computed in O(n+m) time. An unweighted convex bipartite graph has a representation of size O(n) that records for each vertex u∈U the first and last neighbor in the ordering of V. Given such a compact representation, we compute an induced matching of maximum cardinality in O(n) time. In convex bipartite graphs, maximum-cardinality induced matchings are dual to minimum chain covers. A chain cover is a covering of the edge set by chain subgraphs, that is, subgraphs that do not contain induced matchings of more than one edge. Given a compact representation, we compute a representation of a minimum chain cover in O(n) time. If no compact representation is given, the cover can be computed in O(n+m) time. All of our algorithms achieve optimal linear running time for the respective problem and model, and they improve and generalize the previous results in several ways: The best algorithms for the unweighted problem versions had a running time of O(n\(^{2}\)) (Brandstädt et al. in Theor. Comput. Sci. 381(1–3):260–265, 2007. https://doi.org/10.1016/j.tcs.2007.04.006). The weighted case has not been considered before.
Proximity dimensions and the emergence of collaboration: a HypTrails study on German AI research
(2021)
Creation and exchange of knowledge depends on collaboration. Recent work has suggested that the emergence of collaboration frequently relies on geographic proximity. However, being co-located tends to be associated with other dimensions of proximity, such as social ties or a shared organizational environment. To account for such factors, multiple dimensions of proximity have been proposed, including cognitive, institutional, organizational, social and geographical proximity. Since they strongly interrelate, disentangling these dimensions and their respective impact on collaboration is challenging. To address this issue, we propose various methods for measuring different dimensions of proximity. We then present an approach to compare and rank them with respect to the extent to which they indicate co-publications and co-inventions. We adapt the HypTrails approach, which was originally developed to explain human navigation, to co-author and co-inventor graphs. We evaluate this approach on a subset of the German research community, specifically academic authors and inventors active in research on artificial intelligence (AI). We find that social proximity and cognitive proximity are more important for the emergence of collaboration than geographic proximity.
Complexity and Partitions
(2001)
Computational complexity theory usually investigates the complexity of sets, i.e., the complexity of partitions into two parts. But often it is more appropriate to represent natural problems by partitions into more than two parts. A particularly interesting class of such problems consists of classification problems for relations. For instance, a binary relation R typically defines a partitioning of the set of all pairs (x,y) into four parts, classifiable according to the cases where R(x,y) and R(y,x) hold, only R(x,y) or only R(y,x) holds or even neither R(x,y) nor R(y,x) is true. By means of concrete classification problems such as Graph Embedding or Entailment (for propositional logic), this thesis systematically develops tools, in shape of the boolean hierarchy of NP-partitions and its refinements, for the qualitative analysis of the complexity of partitions generated by NP-relations. The Boolean hierarchy of NP-partitions is introduced as a generalization of the well-known and well-studied Boolean hierarchy (of sets) over NP. Whereas the latter hierarchy has a very simple structure, the situation is much more complicated for the case of partitions into at least three parts. To get an idea of this hierarchy, alternative descriptions of the partition classes are given in terms of finite, labeled lattices. Based on these characterizations the Embedding Conjecture is established providing the complete information on the structure of the hierarchy. This conjecture is supported by several results. A natural extension of the Boolean hierarchy of NP-partitions emerges from the lattice-characterization of its classes by considering partition classes generated by finite, labeled posets. It turns out that all significant ideas translate from the case of lattices. The induced refined Boolean hierarchy of NP-partitions enables us more accuratly capturing the complexity of certain relations (such as Graph Embedding) and a description of projectively closed partition classes.
This technical report introduces the Descartes Modeling Language (DML), a new architecture-level modeling language for modeling Quality-of-Service (QoS) and resource management related aspects of modern dynamic IT systems, infrastructures and services. DML is designed to serve as a basis for self-aware resource management during operation ensuring that system QoS requirements are continuously satisfied while infrastructure resources are utilized as efficiently as possible.