004 Datenverarbeitung; Informatik
Refine
Has Fulltext
- yes (203)
Year of publication
Document Type
- Journal article (78)
- Doctoral Thesis (75)
- Working Paper (37)
- Conference Proceeding (8)
- Master Thesis (3)
- Report (2)
Language
- English (183)
- German (19)
- Multiple languages (1)
Keywords
- Datennetz (14)
- Leistungsbewertung (13)
- virtual reality (12)
- Robotik (8)
- Mobiler Roboter (7)
- Autonomer Roboter (6)
- Komplexitätstheorie (5)
- Optimierung (5)
- P4 (5)
- Simulation (5)
- Theoretische Informatik (5)
- artificial intelligence (5)
- machine learning (5)
- Drahtloses Sensorsystem (4)
- Modellierung (4)
- Netzwerk (4)
- Optimization (4)
- Overlay-Netz (4)
- Routing (4)
- SDN (4)
- Verteiltes System (4)
- Algorithmus (3)
- Approximationsalgorithmus (3)
- Computer Vision (3)
- Computersimulation (3)
- Dienstgüte (3)
- Graph (3)
- Graphenzeichnen (3)
- Komplexität (3)
- Künstliche Intelligenz (3)
- Lokalisation (3)
- Mensch-Maschine-Schnittstelle (3)
- Netzwerkmanagement (3)
- Peer-to-Peer-Netz (3)
- Performance Evaluation (3)
- QoE (3)
- Quadrocopter (3)
- Quality of Experience (3)
- Rechnernetz (3)
- Ressourcenmanagement (3)
- Robotics (3)
- approximation algorithm (3)
- augmented reality (3)
- graph drawing (3)
- immersion (3)
- simulation (3)
- 5G (2)
- Algorithmische Geometrie (2)
- Ausfallsicheres System (2)
- Ausfallsicherheit (2)
- Benutzerschnittstelle (2)
- Crowdsourcing (2)
- Data Mining (2)
- Deep learning (2)
- Dot-Depth Problem (2)
- Drahtloses lokales Netz (2)
- Effizienter Algorithmus (2)
- Entscheidbarkeit (2)
- Entscheidungsfindung (2)
- Fernwartung (2)
- Future Internet (2)
- Human-Robot-Interaction (2)
- IEEE 802.11 (2)
- Internet of Things (2)
- IoT (2)
- Kleinsatellit (2)
- Kreuzung (2)
- Localization (2)
- MP-DCCP (2)
- Machine Learning (2)
- Maschinelles Lernen (2)
- Maschinelles Sehen (2)
- Mehragentensystem (2)
- Mensch-Maschine-System (2)
- Mensch-Roboter-Interaktion (2)
- Mixed Reality (2)
- Mustererkennung (2)
- NP-hardness (2)
- PROLOG <Programmiersprache> (2)
- Punktwolke (2)
- Quadrotor (2)
- Resilience (2)
- Resource Management (2)
- Satellit (2)
- Situation Awareness (2)
- Software Defined Networking (2)
- Software Engineering (2)
- Teleoperation (2)
- Theoretical Computer Science (2)
- User Interface (2)
- Verbotsmuster (2)
- XR (2)
- automation (2)
- connected mobility applications (2)
- crossing minimization (2)
- decidability (2)
- deep learning (2)
- dot-depth problem (2)
- education (2)
- educational tool (2)
- endliche Automaten (2)
- endoscopy (2)
- exposure (2)
- finite automata (2)
- fog computing (2)
- forbidden patterns (2)
- fully convolutional neural networks (2)
- games (2)
- gastroenterology (2)
- historical document analysis (2)
- human-computer interaction (2)
- knowledge representation (2)
- measurements (2)
- mobile networks (2)
- mobile robots (2)
- multipath (2)
- multipath scheduling (2)
- navigation (2)
- network calculus (2)
- neural networks (2)
- ontology (2)
- prediction (2)
- regular languages (2)
- reguläre Sprachen (2)
- satellite communication (2)
- scheduling (2)
- self-aware computing (2)
- virtual environments (2)
- 3D Laser Scanning (1)
- 3D Pointcloud (1)
- 3D Punktwolke (1)
- 3D Reconstruction (1)
- 3D Sensor (1)
- 3D Vision (1)
- 3D point cloud (1)
- 3D thermal mapping (1)
- 3D-Rekonstruktion (1)
- 3D-reconstruction methods (1)
- 3DTK toolkit (1)
- 4D-GIS (1)
- 4G Networks (1)
- 5G core network (1)
- 5G-ATSSS (1)
- 5GC (1)
- 6DOF Pose Estimation (1)
- 6G (1)
- ATSSSS (1)
- AVA (1)
- Abhängigskeitsgraph (1)
- Ablaufplanung (1)
- Add-on-Miss (1)
- Admission Control (1)
- Agent <Informatik> (1)
- Agent <Künstliche Intelligenz> (1)
- Agent-based Simulation (1)
- Agentbased System (1)
- Agenten-basierte Simulation (1)
- Angewandte Informatik (1)
- Annotation (1)
- Anwendung (1)
- Anwendungsfall (1)
- Approximation (1)
- Arterie (1)
- Artery (1)
- Aufsatzsammlung (1)
- Aufwandsanalyse (1)
- Automat <Automatentheorie> (1)
- Automata Theory (1)
- Automatentheorie (1)
- Automatisierte Prüfungskorrektur (1)
- Autonomer Agent (1)
- Autonomie (1)
- Autonomous Robot (1)
- Autonomous UAV (1)
- Autonomous multi-vehicle systems (1)
- Autorotation (1)
- Backbone-Netz (1)
- Background Knowledge (1)
- Balloon (1)
- Benutzerinteraktion (1)
- Berechenbarkeit (1)
- Berechnungskomplexität (1)
- Bernoulli stochastics (1)
- Bernoulli-Raum (1)
- Bernoullische Stochastik (1)
- Bernoullispace (1)
- Betriebssystem (1)
- Bewegungsablauf (1)
- Bewegungskompensation (1)
- Bewegungskoordination (1)
- Bewegungsplanung (1)
- Bildverarbeitung (1)
- Bit Parallelität (1)
- BitTorrent (1)
- Bodenstation (1)
- Boolean Grammar (1)
- Boolean equivalence (1)
- Boolean functions (1)
- Boolean hierarchy (1)
- Boolean isomorphism (1)
- Boolesche Funktionen (1)
- Boolesche Grammatik (1)
- Boolesche Hierarchie (1)
- Brüder Grimm Privatbibliothek (1)
- Business Intelligence (1)
- CASE (1)
- CLIP (1)
- Calibration (1)
- Call Graph (1)
- Chord (1)
- Clones (1)
- Cloud Gaming (1)
- Communication Networks (1)
- Complexity Theory (1)
- Compression (1)
- Computational Geometry (1)
- Computational complexity (1)
- Computerspiel (1)
- Computerunterstütztes Lernen (1)
- Containerization (1)
- Content Distribution (1)
- Convolutional Neural Network (1)
- Cost Analysis (1)
- Crowdsensing (1)
- CubeSat (1)
- DHT (1)
- DNA storage (1)
- Daedalus-Projekt (1)
- Databases (1)
- Datenbanken (1)
- Datenbasis (1)
- Datenkommunikationsnetz (1)
- Deep Georeferencing (1)
- Deep Learning (1)
- Dependency Graph (1)
- Design (1)
- Design and Development (1)
- Dezentrale Regelung (1)
- Diagnosesystem (1)
- Dichotomy (1)
- Digitalisierung (1)
- Dijkstra’s algorithm (1)
- Diskrete Simulation (1)
- Distributed Space Systems (1)
- Domänenspezifische Sprache (1)
- Dot-Depth-Hierarchie (1)
- Drahtloses vermaschtes Netz (1)
- Dreidimensionale Rekonstruktion (1)
- Dynamic Environments (1)
- Dynamic Memory Management (1)
- Dynamische Speicherverwaltung (1)
- EPM (1)
- Echtzeitsystem (1)
- Echzeit (1)
- Edge-MEC-Cloud (1)
- Edge-based Intelligence (1)
- Educational Measurement (I2.399) (1)
- Eingebettetes System (1)
- Elasticity tensor (1)
- Elastizitätstensor (1)
- Embedded Systems (1)
- Endnutzer (1)
- Endpoint Mobility (1)
- Energieeffizienz (1)
- Energy efficiency (1)
- Entscheidungsträger (1)
- Erfüllbarkeitsproblem (1)
- Erkennung handschriftlicher Artefakte (1)
- Erweiterte Realität (1)
- Ethik (1)
- Euclidean plane (1)
- Euklidische Ebene (1)
- Expert System (1)
- Expertensystem (1)
- FIFO caching strategies (1)
- Fachgespräch (1)
- Fahrsimulation (1)
- Fahrsimulator (1)
- Fairness (1)
- Fallstudie (1)
- Feature Based Registration (1)
- Feature-Matching (1)
- Fehlertoleranz (1)
- Feldprogrammierbare Architekturen (1)
- Fernsteuerung (1)
- Field programmable gate array (1)
- Field-programmable Gate Arrays (1)
- Flugkörper (1)
- Forces (1)
- Formale Sprache (1)
- Formation (1)
- Formation Flight (1)
- Formationsbewegung (1)
- Forschung (1)
- Frühdruck (1)
- Funkressourcenverwaltung (1)
- GNSS/INS integrated navigation (1)
- Gastroenterologische Endoskopie (1)
- Gay-Array-Bauelement (1)
- Generalisierung <Kartografie> (1)
- Generation Problem (1)
- Generierungsproblem (1)
- Genetic Optimization (1)
- Genetische Optimierung (1)
- Georeferenzierung (1)
- Gllobal self-localisation (1)
- Globale Selbstlokalisation (1)
- Grimm brothers personal library (1)
- Ground Station Networks (1)
- H.264 SVC (1)
- H.264/SVC (1)
- HMD (Head-Mounted Display) (1)
- HSPA (1)
- HTTP adaptive video streaming (1)
- Halbordnungen (1)
- Handschrift (1)
- Hardware (1)
- Herzkatheter (1)
- Herzkathetereingriff (1)
- Hierarchische Simulation (1)
- Hintergrundwissen (1)
- Historical Maps (1)
- Historische Karte (1)
- Historische Landkarten (1)
- Hospital (1)
- IEEE 802.11e (1)
- IEEE 802.15.4 (1)
- INS/LIDAR integrated navigation (1)
- IP (1)
- IT security (1)
- Image Aesthetic Assessment (1)
- Implementierung <Informatik> (1)
- Industrial internet (1)
- Industrie 4.0 (1)
- Inferenz <Künstliche Intelligenz> (1)
- Informatik (1)
- Information Extraction (1)
- Information Retrieval (1)
- Information-Retrieval-System (1)
- Instrument Control Toolbox (1)
- Intelligent Virtual Agents (1)
- InteractionSuitcase (1)
- Interaktion (1)
- Internet (1)
- Internet Protokoll (1)
- Invertierte Liste (1)
- Isomorphie (1)
- Itinerare (1)
- Itineraries (1)
- JCAS (1)
- Java <Programmiersprache> (1)
- Java Message Service (1)
- Kademlia (1)
- Kanalzugriff (1)
- Karte (1)
- Kartierung (1)
- Kathará (1)
- Kerneldensity estimation (1)
- Klassendiagramm (1)
- Klima (1)
- Knowledge Discovery (1)
- Knowledge-based Systems Engineering (1)
- Kombinatorik (1)
- Kommunikation (1)
- Kommunikationsnetze (1)
- Komplexitätsklasse (1)
- Komplexitätsklasse NP (1)
- Konvexe Zeichnungen (1)
- Konzeptsuche (1)
- Kooperierende mobile Roboter (1)
- Krankenhaus (1)
- Kreuzungsminimierung (1)
- Kurve (1)
- LFU (1)
- LRU (1)
- Landkartenbeschriftung (1)
- Landnutzungskartierung (1)
- Laser scanning (1)
- Lava (1)
- Learning (1)
- Lehre (1)
- Lernen (1)
- Lidar (1)
- Lightning (1)
- Link rate adaptation (1)
- Linkratenanpassung (1)
- Linux (1)
- LoRa (1)
- LoRaWAN (1)
- LoRaWan (1)
- Logic Programming (1)
- Logische Programmierung (1)
- Lunar Caves (1)
- Lunar Exploration (1)
- MAC (1)
- MTC (1)
- Mapping (1)
- Mars (1)
- Mathematische Modellierung (1)
- Mathematisches Modell (1)
- Medical Image Analysis (1)
- Medium <Physik> (1)
- Medizin (1)
- Mehrebenensimulation (1)
- Mehrfahrzeugsysteme (1)
- Mehrkriterielle Optimierung (1)
- Mehrpfadübertragung (1)
- Mehrschichtnetze (1)
- Mehrschichtsystem (1)
- Mensch-Maschine-Kommunikation (1)
- Mesh Networks (1)
- Mesh Netze (1)
- Methodologie (1)
- Middleware (1)
- Miniaturisierung (1)
- Minimally invasive vascular intervention (1)
- Missionsbetrieb (1)
- Mobile Roboter (1)
- Mobiles Internet (1)
- Mobilfunk (1)
- Modell (1)
- Modellbasierte Diagnose (1)
- Modellierungstechniken (1)
- Modelling (1)
- Modularität (1)
- Mond (1)
- Motion Planning (1)
- Multi-Agent-Simulation (1)
- Multi-Layer (1)
- Multi-Network Service (1)
- Multi-Netzwerk Dienste (1)
- Multi-Paradigm Programming (1)
- Multi-Paradigm Programming Framework (1)
- Multi-agent system (1)
- Multiagentensimulation (1)
- Multiagentensystem (1)
- Multimedia (1)
- Multipath Transmission (1)
- Multiple-Choice Examination (1)
- Multiple-Choice Prüfungen (1)
- NP (1)
- NP-Vollständigkeit (1)
- NP-complete sets (1)
- NP-hartes Problem (1)
- NP-schweres Problem (1)
- Network Emulator (1)
- Network Experiments (1)
- Network Management (1)
- Network Measurements (1)
- Network Virtualization (1)
- Netzplantechnik (1)
- Netzplanung (1)
- Netzvirtualisierung (1)
- Netzwerkplanung (1)
- Netzwerkvirtualisierung (1)
- Neuronales Netz (1)
- Next Generation Networks (1)
- Nichtholonome Fahrzeuge (1)
- Nichtlineare Regelung (1)
- Object Detection (1)
- Object-Oriented Programming (1)
- Objektorientierte Programmierung (1)
- Open Source (1)
- Operator (1)
- Optical Flow (1)
- Optimale Kontrolle (1)
- Optimierungsproblem (1)
- Overlay (1)
- Overlay Netzwerke (1)
- Overlay networks (1)
- Overlays (1)
- P4-INT (1)
- Panorama Images (1)
- Parameterkalibrierung (1)
- Partition <Mengenlehre> (1)
- Partitionen (1)
- Path Computation Element (1)
- Pattern Recognition (1)
- Peer-to-Peer (1)
- Performance Analysis (1)
- Performance Enhancing Proxies (1)
- Performance Management (1)
- Performance Modeling (1)
- Pfadberechnungselement (1)
- Picosatellite (1)
- Planare Graphen (1)
- Planausführung (1)
- Planung (1)
- Planungssystem (1)
- Poisson surface reconstruction (1)
- Polyeder (1)
- Polypektomie (1)
- Post's Classes (1)
- Postsche Klassen (1)
- Prediction (1)
- Process Optimization (1)
- Programmierbare logische Anordnung (1)
- Prozessoptimierung (1)
- Publish-Subscribe-System (1)
- QUIC (1)
- QoS (1)
- Quality of Experience (QoE) (1)
- Quality of Experience QoE (1)
- Quality of Service (1)
- Quality of Service (QoS) (1)
- Quality-of-Experience (1)
- Quality-of-Service (1)
- Quality-of-Service (QoS) (1)
- Quantor (1)
- RGB-D (1)
- RRM (1)
- Raumdaten (1)
- Real-Time Operating Systems (1)
- Real-time (1)
- Refactoring (1)
- Regelbasiertes System (1)
- Regelung (1)
- Registration (1)
- Registrierung (1)
- Reguläre Sprache (1)
- Rendezvous (1)
- Reproducibility (1)
- Resource and Performance Management (1)
- Ressourcen Management (1)
- Ressourcenallokation (1)
- Rettungsroboter (1)
- Robot (1)
- Roboter (1)
- Rule-based Systems (1)
- SBA (1)
- SDN/NVF (1)
- Scheduling (1)
- Search-and-Rescue (1)
- Selbstkalibrierung (1)
- Self-Evaluation Programs (I2.399.780) (1)
- Self-calibration (1)
- Semantic Web (1)
- Semantics (1)
- Semantik (1)
- Sensing-aaS (1)
- Sensor (1)
- Service Mobility (1)
- Sichtbarkeit (1)
- Similarity Measure (1)
- Simulator (1)
- Situationsbewusstsein (1)
- Skype (1)
- Small Satellites (1)
- Smart User Interaction (1)
- Social Web (1)
- Software (1)
- Software Performance Engineering (1)
- Software Performance Modeling (1)
- Source Code Visualization (1)
- Soziale Software (1)
- Spherical Robot (1)
- Standardisierung (1)
- Standortproblem (1)
- Sternfreie Sprache (1)
- Steuerung (1)
- Stiffness (1)
- Stochastik (1)
- Stochastikon (1)
- Strahlentherapie (1)
- Straubing-Th´erien-Hierarchie (1)
- Straßennetzwerk (1)
- Straßenverkehr (1)
- Structure-from-Motion (1)
- Strukturelle Komplexität (1)
- Subgroup Mining (1)
- Subgruppenentdeckung (1)
- Suchverfahren (1)
- System (1)
- Szenariogenerierung (1)
- TSN (1)
- TTL (1)
- TTL validation of data consistency (1)
- Teaching (1)
- Telematik (1)
- Terramechanics (1)
- Testbed (1)
- Theoretical computer science (1)
- Thermografie (1)
- Torque (1)
- Trainingssystem (1)
- Travelling-salesman-Problem (1)
- Tumor motion (1)
- Tumorbewegung (1)
- U-Bahnlinienplan (1)
- UI and Interaction Design (1)
- UML Klassendiagramm (1)
- UML class diagram (1)
- UMTS (1)
- URL (1)
- URLLC (1)
- Underwater Mapping (1)
- Underwater Scanning (1)
- Unmanned Aerial Vehicle (1)
- Unstetige Regelung (1)
- Usability (1)
- Use case (1)
- User Behavior (1)
- User Participation (1)
- VNF (1)
- VPN (1)
- Venus (1)
- Veranstaltung (1)
- Verbotenes Muster (1)
- Verbände (1)
- Verkehrslenkung (1)
- Verteilung von Inhalten (1)
- Video Quality Monitoring (1)
- Video Streaming (1)
- Videoübertragung (1)
- Virtualisierung (1)
- Virtuelles Netzwerk (1)
- Visibility (1)
- Visual Tracking (1)
- Visualisierung (1)
- Visualization (1)
- Visualized Kathará (1)
- Voice-over-IP (VoIP) (1)
- Volltextsuche (1)
- Vorhersage (1)
- WLAN (1)
- Warteschlangentheorie (1)
- Werkstattdiagnose (1)
- WhatsApp (1)
- Wheel (1)
- Winkel (1)
- Wire relaxation (1)
- Wireless LAN (1)
- Wireless Sensor/Actuator Systems (1)
- Wissensbasiertes System (1)
- Wissensendeckung (1)
- Wissensrepräsentation (1)
- Worterweiterungen (1)
- XR-artificial intelligence combination (1)
- XR-artificial intelligence continuum (1)
- YouTube (1)
- Zeichnen von Graphen (1)
- Zeitdiskretes System (1)
- Zugangskontrolle (1)
- Zählprobleme (1)
- abgeschlossene Klassen (1)
- acrophobia (1)
- adaptation models (1)
- administrative boundary (1)
- admission control (1)
- adult learning (1)
- aerodynamics (1)
- aerospace (1)
- aftermarket diagnostic (1)
- agent-based models (1)
- agents (1)
- agile Prozesse (1)
- agile processes (1)
- angular schematization (1)
- annotation (1)
- anomaly detection (1)
- anomaly prediction (1)
- ant-colony optimization (1)
- anthropomorphism (1)
- anxiety (1)
- application design (1)
- approximation algorithms (1)
- arithmetic calculations (1)
- automatic Layout (1)
- automatisches Layout (1)
- autonomic orchestration (1)
- autonomous (1)
- autonomous UAV (1)
- autorotation (1)
- availability (1)
- avatar embodiment (1)
- avatars (1)
- background knowledge (1)
- baseline detection (1)
- behavior perception (1)
- binary tanglegram (1)
- biosignals (1)
- bit-parallel (1)
- boundary labeling (1)
- building (1)
- camera orientation (1)
- car-like robots (1)
- carbon (1)
- case study (1)
- certifying algorithm (1)
- chain cover (1)
- channel management (1)
- climate (1)
- co-authorships (1)
- co-inventorships (1)
- coherence (1)
- collaboration (1)
- collision (1)
- communication models (1)
- communication networks (1)
- competitive location (1)
- complexity (1)
- computational complexity (1)
- computer performance evaluation (1)
- computergestützte Softwaretechnik (1)
- concept search (1)
- congruence (1)
- constrained forest (1)
- contact representation (1)
- container virtualization (1)
- content-based image retrieval (1)
- continuous-time SLAM (1)
- convex bipartite graph (1)
- convolutional neural network (1)
- cost-sensitive learning (1)
- counting problems (1)
- crowdsensing (1)
- crowdsourced QoE measurements (1)
- crowdsourced measurements (1)
- crowdsourced network measurements (1)
- cultural and media studies (1)
- culturally aware (1)
- curves (1)
- d3web.Train (1)
- data mining (1)
- data plane programming (1)
- data structure (1)
- data warehouse (1)
- database (1)
- dataplane programming (1)
- decision support system (1)
- decision-making (1)
- decission finding (1)
- decoding error rate (1)
- deep metric learning (1)
- definite clause grammars (1)
- deformation-based method (1)
- delay QoS exponent (1)
- delay bound violation probability (1)
- delay constrained (1)
- descent (1)
- design (1)
- design cycle (1)
- detection time simulation (1)
- dial a ride (1)
- digital twin (1)
- dimensions of proximity (1)
- discrete-time analysis (1)
- disjoint multi-paths (1)
- distributed control (1)
- docker (1)
- driving simulation (1)
- dynamic flow migration (1)
- dynamic programming (1)
- dynamische Umgebungen (1)
- eHealth (1)
- early printed books (1)
- edge labeled graphs (1)
- effective Bandwidth (1)
- efficient algorithm (1)
- electronic health records (1)
- elevated plus-maze (1)
- embedding techniques (1)
- emotions (1)
- emulation (1)
- encryption (1)
- end user (1)
- endurance (1)
- energy efficiency (1)
- event detection (1)
- exercise intensity (1)
- experience (1)
- experimental evaluation (1)
- extended reality (XR) (1)
- failure prediction (1)
- fast reroute (1)
- fault detection (1)
- feature matching (1)
- feature-matching (1)
- federated learning (1)
- few-shot learning (1)
- field-programmable architectures (1)
- field-programmable gate arrays (1)
- fixed-parameter tractability (1)
- food quality (1)
- foreign language learning and teaching (1)
- formation driving (1)
- formation flight (1)
- fruit temperature (1)
- full-text search (1)
- future Internet architecture (1)
- future energy grid exploration (1)
- game mechanics (1)
- gamification (1)
- genetic algorithm (1)
- global IPX network (1)
- graph (1)
- graph algorithm (1)
- graph decomposition (1)
- graphs (1)
- group-based communication (1)
- handwriting (1)
- handwritten artefact recognition (1)
- haptic data (1)
- hardness (1)
- hardware-in-the-loop simulation (1)
- hardware-in-the-loop streaming system (1)
- hierarchy (1)
- historical images (1)
- hit ratio analysis and simulation (1)
- hospital data (1)
- human body weight (1)
- human computer interaction (HCI) (1)
- human-artificial intelligence interaction (1)
- human-artificial intelligence interface (1)
- human-centered design (1)
- human-centered, human-robot (1)
- human–computer interaction (1)
- hybrid Diagnostic (1)
- hybrid access (1)
- hybride Diagnose (1)
- illusion of self-motion (1)
- image classification (1)
- image processing (1)
- imbalanced regression (1)
- immersive classroom (1)
- immersive classroom management (1)
- immersive learning technologies (1)
- immersive technologies (1)
- implicit association test (1)
- independent crossing (1)
- induced matching (1)
- informal education (1)
- information extraction (1)
- information systems and information technology (1)
- intelligent transportation systems (1)
- intelligent vehicles (1)
- intelligent virtual agents (1)
- intelligent voice assistant (1)
- intelligente Applikationen (1)
- interactive authoring system (1)
- intercultural learning and teaching (1)
- interdisciplinary education (1)
- internet protocol (1)
- internet traffic (1)
- invasive vascular interventions (1)
- iowa gambling task (1)
- key-insight extraction (1)
- kinect (1)
- labeling (1)
- land-cover area (1)
- landing (1)
- language-image pre-training (1)
- latency (1)
- lattices (1)
- layout recognition (1)
- learning environments (1)
- least cost (1)
- light-gated proteins (1)
- load balancing (1)
- local energy system (1)
- locomotion (1)
- logic programming (1)
- logistics (1)
- long-term analysis (1)
- map projections (1)
- mapping (1)
- mathematical model (1)
- media analysis (1)
- medical records (1)
- medieval manuscripts (1)
- meditation (1)
- metro map (1)
- mindfulness (1)
- misconceptions (1)
- mission operation (1)
- mixed reality (1)
- mixed-cultural (1)
- mixed-cultural settings (1)
- mobile instant messaging (1)
- mobile messaging application (1)
- mobile streaming (1)
- model following (1)
- model output statistics (1)
- model predictive control (1)
- model-base diagnosis (1)
- model-based diagnosis (1)
- modeling techniques (1)
- monotone drawing (1)
- morphing (1)
- multi-vehicle formations (1)
- multi-vehicle rendezvous (1)
- multimodal fusion (1)
- multimodal interface (1)
- multimodal learning (1)
- multipath communication (1)
- multipath packet scheduling (1)
- multiple myeloma (1)
- multirotors (1)
- multiscale encoder (1)
- n-Gramm (1)
- n-gram (1)
- nano-satellite (1)
- nanocellulose (1)
- natural interfaces (1)
- natural language processing (1)
- natural language processing · · · (1)
- natural user interfaces (1)
- network (1)
- network design (1)
- network planning (1)
- network softwarization (1)
- network upgrade (1)
- network virtualization (1)
- networked robotics (1)
- networking (1)
- networks (1)
- neume notation (1)
- neural architecture (1)
- non-native accent (1)
- non-terrestrial networks (1)
- nonholonomic vehicles (1)
- object detection (1)
- octree (1)
- open source (1)
- optical music recognition (1)
- optimization (1)
- orchestration (1)
- overprovisioning (1)
- packet reception method (1)
- partitions (1)
- passage of time (1)
- passive haptic feedback (1)
- path computation (1)
- pattern perception (1)
- perception (1)
- performance (1)
- performance analysis (1)
- performance evaluation (1)
- performance monitoring (1)
- performance prediction (1)
- place-illusion (1)
- plain orchestrating service (1)
- plan execution (1)
- plausibility (1)
- plausibility-illusion (1)
- point cloud (1)
- point cloud compression (1)
- point-to-plane measure (1)
- point-to-point measure (1)
- pollution (1)
- pos (1)
- posets (1)
- positioning (1)
- precision horticulture (1)
- precision training (1)
- presence (1)
- private chat groups (1)
- procedural content generation (1)
- procedural fusion methods (1)
- process model (1)
- prompt engineering (1)
- protein chip (1)
- psychomotor training (1)
- psychophyisology (1)
- public speaking (1)
- q-Gramm (1)
- q-gram (1)
- quadcopter (1)
- quadcopters (1)
- quality assurance (1)
- quality evaluation (1)
- quality of experience (1)
- quality of experience prediction (1)
- radio resource management (1)
- radiology (1)
- ransomware (1)
- real world evidence (1)
- real-world application (1)
- realism (1)
- receding horizon control (1)
- recommender system (1)
- reconfiguration (1)
- regelbasierte Nachbearbeitung (1)
- reload cost (1)
- remote control (1)
- research methods (1)
- resilience (1)
- rich vehicle routing problem (1)
- right angle crossing (1)
- road network (1)
- robotics (1)
- robustness (1)
- rotorcraft (1)
- rotors (1)
- routing (1)
- rule based post processing (1)
- sample weighting (1)
- satisfiability problems (1)
- scalability (1)
- scalability evaluation (1)
- scalable quadcopter (1)
- scenario creation (1)
- science, technology and society (1)
- secure group communication (1)
- segmentation (1)
- self-adaptive (1)
- self-adaptive systems (1)
- self-assembly (1)
- self-aware computing systems (1)
- self-managing systems (1)
- semantic fusion (1)
- semantic understanding (1)
- semantic web (1)
- semantical aesthetic (1)
- semantische Ästhetik (1)
- sensor (1)
- sensor devices (1)
- sensor fusion (1)
- sensor network (1)
- sentinel (1)
- serious games (1)
- service-curve estimation (1)
- sesnsors (1)
- short block-length (1)
- shortest path routing (1)
- signaling traffic (1)
- simulation system (1)
- simultaneous embedding (1)
- single-electron transistors (1)
- skalierbare Diagnose (1)
- sketching (1)
- slam (1)
- smart meter data utilization (1)
- smart speaker (1)
- smooth orthogonal drawing (1)
- snow shoveling (1)
- social VR (1)
- social interaction (1)
- social relationship (1)
- social robot (1)
- social robotics (1)
- social role (1)
- socially interactive agents (1)
- software defined network (1)
- spanning tree (1)
- spatial presence (1)
- standardization (1)
- state management (1)
- statistical validity (1)
- statistics and numerical data (1)
- stereotypes (1)
- stochastic thinking (1)
- stochastisches Denken (1)
- stroke (1)
- structural complexity (1)
- student simulation (1)
- stylus (1)
- sunburn (1)
- supervised learning (1)
- surface model (1)
- survey (1)
- sustainability (1)
- switching navigation (1)
- system (1)
- systematic literature review (1)
- systematic review (1)
- table extraction (1)
- table understanding (1)
- taxonomy (1)
- teacher education (1)
- technology-supported learning (1)
- telematics (1)
- text line detection (1)
- text supervision (1)
- theory (1)
- therapeutic application (1)
- thermal camera (1)
- thermal point cloud (1)
- time calibration (1)
- time perception (1)
- timestamping method (1)
- tools (1)
- topology (1)
- training systems (1)
- trait anxiety (1)
- trajectory planning (1)
- transformer (1)
- translational neuroscience (1)
- transport microenvironments (1)
- transport protocols (1)
- transportation (1)
- tree (1)
- unmanned aerial vehicle (1)
- unmanned aerial vehicles (1)
- usability evaluation (1)
- use cases (1)
- user experience (1)
- user interaction (1)
- user study (1)
- user-generated content (1)
- v (1)
- vection (1)
- vehicle dynamics (1)
- vehicular navigation (1)
- verbal behaviour (1)
- vernetzte Roboter (1)
- virtual agent (1)
- virtual agent interaction (1)
- virtual humans (1)
- virtual queue (1)
- virtual reality training (1)
- virtual stimuli (1)
- virtual tunnel (1)
- virtual-reality-continuum (1)
- vom Nutzer erfahrene Dienstgüte QoE (1)
- voting location (1)
- waypoint parameter (1)
- wearable (1)
- wheel (1)
- wireless network (1)
- word clouds (1)
- word extensions (1)
- zooming (1)
- zukünftige Kommunikationsnetze (1)
- zukünftiges Internet (1)
- Ähnlichkeitsmaß (1)
Institute
- Institut für Informatik (203) (remove)
Schriftenreihe
Sonstige beteiligte Institutionen
Utilizing multiple access technologies such as 5G, 4G, and Wi-Fi within a coherent framework is currently standardized by 3GPP within 5G ATSSS. Indeed, distributing packets over multiple networks can lead to increased robustness, resiliency and capacity. A key part of such a framework is the multi-access proxy, which transparently distributes packets over multiple paths. As the proxy needs to serve thousands of customers, scalability and performance are crucial for operator deployments. In this paper, we leverage recent advancements in data plane programming, implement a multi-access proxy based on the MP-DCCP tunneling approach in P4 and hardware accelerate it by deploying the pipeline on a smartNIC. This is challenging due to the complex scheduling and congestion control operations involved. We present our pipeline and data structures design for congestion control and packet scheduling state management. Initial measurements in our testbed show that packet latency is in the range of 25 μs demonstrating the feasibility of our approach.
An innovative framework has been developed for teamwork of two quadcopter formations, each having its specified formation geometry, assigned task, and matching control scheme. Position control for quadcopters in one of the formations has been implemented through a Linear Quadratic Regulator Proportional Integral (LQR PI) control scheme based on explicit model following scheme. Quadcopters in the other formation are controlled through LQR PI servomechanism control scheme. These two control schemes are compared in terms of their performance and control effort. Both formations are commanded by respective ground stations through virtual leaders. Quadcopters in formations are able to track desired trajectories as well as hovering at desired points for selected time duration. In case of communication loss between ground station and any of the quadcopters, the neighboring quadcopter provides the command data, received from the ground station, to the affected unit. Proposed control schemes have been validated through extensive simulations using MATLAB®/Simulink® that provided favorable results.
A centralized heterogeneous formation flight position control scheme has been formulated using an explicit model following design, based on a Linear Quadratic Regulator Proportional Integral (LQR PI) controller. The leader quadcopter is a stable reference model with desired dynamics whose output is perfectly tracked by the two wingmen quadcopters. The leader itself is controlled through the pole placement control method with desired stability characteristics, while the two followers are controlled through a robust and adaptive LQR PI control method. Selected 3-D formation geometry and static stability are maintained under a number of possible perturbations. With this control scheme, formation geometry may also be switched to any arbitrary shape during flight, provided a suitable collision avoidance mechanism is incorporated. In case of communication loss between the leader and any of the followers, the other follower provides the data, received from the leader, to the affected follower. The stability of the closed-loop system has been analyzed using singular values. The proposed approach for the tightly coupled formation flight of mini unmanned aerial vehicles has been validated with the help of extensive simulations using MATLAB/Simulink, which provided promising results.
A simple test setup has been developed at Institute of Aerospace Information Technology, University of Würzburg, Germany to realize basic functionalities for formation flight of quadrocopters. The test environment is planned to be utilized for developing and validating the algorithms for formation flying capability in real environment as well as for education purpose. An already existing test bed for single quadrocopter was extended with necessary inter-communication and distributed control mechanism to test the algorithms for formation flights in 2 degrees of freedom (roll / pitch). This study encompasses the domain of communication, control engineering and embedded systems programming. Bluetooth protocol has been used for inter-communication between two quadrocopters. A simple approach of PID control in combination with Kalman filter has been exploited. MATLAB Instrument Control Toolbox has been used for data display, plotting and analysis. Plots can be drawn in real-time and received information can also be stored in the form of files for later use and analysis. The test setup has been developed indigenously and at considerably low cost. Emphasis has been placed on simplicity to facilitate students learning process. Several lessons have been learnt during the course of development of this setup. Proposed setup is quite flexible that can be modified as per changing requirements.
While teleoperation of technical highly sophisticated systems has already been a wide field of research, especially for space and robotics applications, the automation industry has not yet benefited from its results. Besides the established fields of application, also production lines with industrial robots and the surrounding plant components are in need of being remotely accessible. This is especially critical for maintenance or if an unexpected problem cannot be solved by the local specialists.
Special machine manufacturers, especially robotics companies, sell their technology worldwide. Some factories, for example in emerging economies, lack qualified personnel for repair and maintenance tasks. When a severe failure occurs, an expert of the manufacturer needs to fly there, which leads to long down times of the machine or even the whole production line. With the development of data networks, a huge part of those travels can be omitted, if appropriate teleoperation equipment is provided.
This thesis describes the development of a telemaintenance system, which was established in an active production line for research purposes. The customer production site of Braun in Marktheidenfeld, a factory which belongs to Procter & Gamble, consists of a six-axis cartesian industrial robot by KUKA Industries, a two-component injection molding system and an assembly unit. The plant produces plastic parts for electric toothbrushes.
In the research projects "MainTelRob" and "Bayern.digital", during which this plant was utilised, the Zentrum für Telematik e.V. (ZfT) and its project partners develop novel technical approaches and procedures for modern telemaintenance. The term "telemaintenance" hereby refers to the integration of computer science and communication technologies into the maintenance strategy. It is particularly interesting for high-grade capital-intensive goods like industrial robots. Typical telemaintenance tasks are for example the analysis of a robot failure or difficult repair operations. The service department of KUKA Industries is responsible for the worldwide distributed customers who own more than one robot. Currently such tasks are offered via phone support and service staff which travels abroad. They want to expand their service activities on telemaintenance and struggle with the high demands of teleoperation especially regarding security infrastructure. In addition, the facility in Marktheidenfeld has to keep up with the high international standards of Procter & Gamble and wants to minimize machine downtimes. Like 71.6 % of all German companies, P&G sees a huge potential for early information on their production system, but complains about the insufficient quality and the lack of currentness of data.
The main research focus of this work lies on the human machine interface for all human tasks in a telemaintenance setup. This thesis provides own work in the use of a mobile device in context of maintenance, describes new tools on asynchronous remote analysis and puts all parts together in an integrated telemaintenance infrastructure. With the help of Augmented Reality, the user performance and satisfaction could be raised. A special regard is put upon the situation awareness of the remote expert realized by different camera viewpoints. In detail the work consists of:
- Support of maintenance tasks with a mobile device
- Development and evaluation of a context-aware inspection tool
- Comparison of a new touch-based mobile robot programming device to the former teach pendant
- Study on Augmented Reality support for repair tasks with a mobile device
- Condition monitoring for a specific plant with industrial robot
- Human computer interaction for remote analysis of a single plant cycle
- A big data analysis tool for a multitude of cycles and similar plants
- 3D process visualization for a specific plant cycle with additional virtual information
- Network architecture in hardware, software and network infrastructure
- Mobile device computer supported collaborative work for telemaintenance
- Motor exchange telemaintenance example in running production environment
- Augmented reality supported remote plant visualization for better situation awareness
Cover contact graphs
(2012)
We study problems that arise in the context of covering certain geometric objects called seeds (e.g., points or disks) by a set of other geometric objects called cover (e.g., a set of disks or homothetic triangles). We insist that the interiors of the seeds and the cover elements are pairwise disjoint, respectively, but they can touch. We call the contact graph of a cover a cover contact graph (CCG). We are interested in three types of tasks, both in the general case and in the special case of seeds on a line: (a) deciding whether a given seed set has a connected CCG, (b) deciding whether a given graph has a realization as a CCG on a given seed set, and (c) bounding the sizes of certain classes of CCG’s. Concerning (a) we give efficient algorithms for the case that seeds are points and show that the problem becomes hard if seeds and covers are disks. Concerning (b) we show that this problem is hard even for point seeds and disk covers (given a fixed correspondence between graph vertices and seeds). Concerning (c) we obtain upper and lower bounds on the number of CCG’s for point seeds.
Data mining has proved its significance in various domains and applications. As an important subfield of the general data mining task, subgroup mining can be used, e.g., for marketing purposes in business domains, or for quality profiling and analysis in medical domains. The goal is to efficiently discover novel, potentially useful and ultimately interesting knowledge. However, in real-world situations these requirements often cannot be fulfilled, e.g., if the applied methods do not scale for large data sets, if too many results are presented to the user, or if many of the discovered patterns are already known to the user. This thesis proposes a combination of several techniques in order to cope with the sketched problems: We discuss automatic methods, including heuristic and exhaustive approaches, and especially present the novel SD-Map algorithm for exhaustive subgroup discovery that is fast and effective. For an interactive approach we describe techniques for subgroup introspection and analysis, and we present advanced visualization methods, e.g., the zoomtable that directly shows the most important parameters of a subgroup and that can be used for optimization and exploration. We also describe various visualizations for subgroup comparison and evaluation in order to support the user during these essential steps. Furthermore, we propose to include possibly available background knowledge that is easy to formalize into the mining process. We can utilize the knowledge in many ways: To focus the search process, to restrict the search space, and ultimately to increase the efficiency of the discovery method. We especially present background knowledge to be applied for filtering the elements of the problem domain, for constructing abstractions, for aggregating values of attributes, and for the post-processing of the discovered set of patterns. Finally, the techniques are combined into a knowledge-intensive process supporting both automatic and interactive methods for subgroup mining. The practical significance of the proposed approach strongly depends on the available tools. We introduce the VIKAMINE system as a highly-integrated environment for knowledge-intensive active subgroup mining. Also, we present an evaluation consisting of two parts: With respect to objective evaluation criteria, i.e., comparing the efficiency and the effectiveness of the subgroup discovery methods, we provide an experimental evaluation using generated data. For that task we present a novel data generator that allows a simple and intuitive specification of the data characteristics. The results of the experimental evaluation indicate that the novel SD-Map method outperforms the other described algorithms using data sets similar to the intended application concerning the efficiency, and also with respect to precision and recall for the heuristic methods. Subjective evaluation criteria include the user acceptance, the benefit of the approach, and the interestingness of the results. We present five case studies utilizing the presented techniques: The approach has been successfully implemented in medical and technical applications using real-world data sets. The method was very well accepted by the users that were able to discover novel, useful, and interesting knowledge.
Operators of Higher Order
(1998)
Motivated by results on interactive proof systems we investigate the computational power of quantifiers applied to well-known complexity classes.
In special, we are interested in existential, universal and probabilistic bounded error quantifiers ranging over words and sets of words, i.e. oracles if we think in a Turing machine model.
In addition to the standard oracle access mechanism, we also consider quantifiers ranging over oracles to which access is restricted in a certain way.
A complete simulation system is proposed that can be used as an educational tool by physicians in training basic skills of Minimally Invasive Vascular Interventions. In the first part, a surface model is developed to assemble arteries having a planar segmentation. It is based on Sweep Surfaces and can be extended to T- and Y-like bifurcations. A continuous force vector field is described, representing the interaction between the catheter and the surface. The computation time of the force field is almost unaffected when the resolution of the artery is increased.
The mechanical properties of arteries play an essential role in the study of the circulatory system dynamics, which has been becoming increasingly important in the treatment of cardiovascular diseases. In Virtual Reality Simulators, it is crucial to have a tissue model that responds in real time. In this work, the arteries are discretized by a two dimensional mesh and the nodes are connected by three kinds of linear springs. Three tissue layers (Intima, Media, Adventitia) are considered and, starting from the stretch-energy density, some of the elasticity tensor components are calculated. The physical model linearizes and homogenizes the material response, but it still contemplates the geometric nonlinearity. In general, if the arterial stretch varies by 1% or less, then the agreement between the linear and nonlinear models is trustworthy.
In the last part, the physical model of the wire proposed by Konings is improved. As a result, a simpler and more stable method is obtained to calculate the equilibrium configuration of the wire. In addition, a geometrical method is developed to perform relaxations. It is particularly useful when the wire is hindered in the physical method because of the boundary conditions. The physical and the geometrical methods are merged, resulting in efficient relaxations. Tests show that the shape of the virtual wire agrees with the experiment. The proposed algorithm allows real-time executions and the hardware to assemble the simulator has a low cost.
In the present work, a simulation system is proposed that can be used as an educational tool by physicians in training basic skills of minimally invasive vascular interventions. In order to accomplish this objective, initially the physical model of the wire proposed by Konings has been improved. As a result, a simpler and more stable method was obtained to calculate the equilibrium configuration of the wire. In addition, a geometrical method is developed to perform relaxations. It is particularly useful when the wire is hindered in the physical method because of the boundary conditions. Then a recipe is given to merge the physical and the geometrical methods, resulting in efficient relaxations. Moreover, tests have shown that the shape of the virtual wire agrees with the experiment. The proposed algorithm allows real-time executions, and furthermore, the hardware to assemble the simulator has a low cost.
Realistic and lifelike 3D-reconstruction of virtual humans has various exciting and important use cases. Our and others’ appearances have notable effects on ourselves and our interaction partners in virtual environments, e.g., on acceptance, preference, trust, believability, behavior (the Proteus effect), and more. Today, multiple approaches for the 3D-reconstruction of virtual humans exist. They significantly vary in terms of the degree of achievable realism, the technical complexities, and finally, the overall reconstruction costs involved. This article compares two 3D-reconstruction approaches with very different hardware requirements. The high-cost solution uses a typical complex and elaborated camera rig consisting of 94 digital single-lens reflex (DSLR) cameras. The recently developed low-cost solution uses a smartphone camera to create videos that capture multiple views of a person. Both methods use photogrammetric reconstruction and template fitting with the same template model and differ in their adaptation to the method-specific input material. Each method generates high-quality virtual humans ready to be processed, animated, and rendered by standard XR simulation and game engines such as Unreal or Unity. We compare the results of the two 3D-reconstruction methods in an immersive virtual environment against each other in a user study. Our results indicate that the virtual humans from the low-cost approach are perceived similarly to those from the high-cost approach regarding the perceived similarity to the original, human-likeness, beauty, and uncanniness, despite significant differences in the objectively measured quality. The perceived feeling of change of the own body was higher for the low-cost virtual humans. Quality differences were perceived more strongly for one’s own body than for other virtual humans.
This work takes a close look at several quite different research areas related to the design of networked embedded sensor/actuator systems. The variety of the topics illustrates the potential complexity of current sensor network applications; especially when enriched with actuators for proactivity and environmental interaction. Besides their conception, development, installation and long-term operation, we'll mainly focus on more "low-level" aspects: Compositional hardware and software design, task cooperation and collaboration, memory management, and real-time operation will be addressed from a local node perspective. In contrast, inter-node synchronization, communication, as well as sensor data acquisition, aggregation, and fusion will be discussed from a rather global network view. The diversity in the concepts was intentionally accepted to finally facilitate the reliable implementation of truly complex systems. In particular, these should go beyond the usual "sense and transmit of sensor data", but show how powerful today's networked sensor/actuator systems can be despite of their low computational performance and constrained hardware: If their resources are only coordinated efficiently!
The steadily increasing usage of smart meters generates a valuable amount of high-resolution data about the individual energy consumption and production of local energy systems. Private households install more and more photovoltaic systems, battery storage and big consumers like heat pumps. Thus, our vision is to augment these collected smart meter time series of a complete system (e.g., a city, town or complex institutions like airports) with simulatively added previously named components. We, therefore, propose a novel digital twin of such an energy system based solely on a complete set of smart meter data including additional building data. Based on the additional geospatial data, the twin is intended to represent the addition of the abovementioned components as realistically as possible. Outputs of the twin can be used as a decision support for either system operators where to strengthen the system or for individual households where and how to install photovoltaic systems and batteries. Meanwhile, the first local energy system operators had such smart meter data of almost all residential consumers for several years. We acquire those of an exemplary operator and discuss a case study presenting some features of our digital twin and highlighting the value of the combination of smart meter and geospatial data.
The development of ICT infrastructures has facilitated the emergence of new paradigms for looking at society and the environment over the last few years. Participatory environmental sensing, i.e. directly involving citizens in environmental monitoring, is one example, which is hoped to encourage learning and enhance awareness of environmental issues. In this paper, an analysis of the behaviour of individuals involved in noise sensing is presented. Citizens have been involved in noise measuring activities through the WideNoise smartphone application. This application has been designed to record both objective (noise samples) and subjective (opinions, feelings) data. The application has been open to be used freely by anyone and has been widely employed worldwide. In addition, several test cases have been organised in European countries. Based on the information submitted by users, an analysis of emerging awareness and learning is performed. The data show that changes in the way the environment is perceived after repeated usage of the application do appear. Specifically, users learn how to recognise different noise levels they are exposed to. Additionally, the subjective data collected indicate an increased user involvement in time and a categorisation effect between pleasant and less pleasant environments.
The rapid development of green and sustainable materials opens up new possibilities in the field of applied research. Such materials include nanocellulose composites that can integrate many components into composites and provide a good chassis for smart devices. In our study, we evaluate four approaches for turning a nanocellulose composite into an information storage or processing device: 1) nanocellulose can be a suitable carrier material and protect information stored in DNA. 2) Nucleotide-processing enzymes (polymerase and exonuclease) can be controlled by light after fusing them with light-gating domains; nucleotide substrate specificity can be changed by mutation or pH change (read-in and read-out of the information). 3) Semiconductors and electronic capabilities can be achieved: we show that nanocellulose is rendered electronic by iodine treatment replacing silicon including microstructures. Nanocellulose semiconductor properties are measured, and the resulting potential including single-electron transistors (SET) and their properties are modeled. Electric current can also be transported by DNA through G-quadruplex DNA molecules; these as well as classical silicon semiconductors can easily be integrated into the nanocellulose composite. 4) To elaborate upon miniaturization and integration for a smart nanocellulose chip device, we demonstrate pH-sensitive dyes in nanocellulose, nanopore creation, and kinase micropatterning on bacterial membranes as well as digital PCR micro-wells. Future application potential includes nano-3D printing and fast molecular processors (e.g., SETs) integrated with DNA storage and conventional electronics. This would also lead to environment-friendly nanocellulose chips for information processing as well as smart nanocellulose composites for biomedical applications and nano-factories.
Diagnostic Case Based Training Systems (D-CBT) provide learners with a means to learn and exercise knowledge in a realistic context. In medical education, D-CBT Systems present virtual patients to the learners who are asked to examine, diagnose and state therapies for these patients. Due a number of conflicting and changing requirements, e.g. time for learning, authoring effort, several systems were developed so far. These systems range from simple, easy-to-use presentation systems to highly complex knowledge based systems supporting explorative learning. This thesis presents an approach and tools to create D-CBT systems from existing sources (documents, e.g. dismissal records) using existing tools (word processors): Authors annotate and extend the documents to model the knowledge. A scalable knowledge representation is able to capture the content on multiple levels, from simple to highly structured knowledge. Thus, authoring of D-CBT systems requires less prerequisites and pre-knowledge and is faster than approaches using specialized authoring environments. Also, authors can iteratively add and structure more knowledge to adapt training cases to their learners needs. The theses also discusses the application of the same approach to other domains, especially to knowledge acquisition for the Semantic Web.
Das stochastische Denken, die Bernoullische Stochastik und dessen informationstechnologische Umsetzung, namens Stochastikon stellen die Grundlage für das Verständnis und die erfolgreiche Nutzung einer stochastischen Wissenschaft dar. Im Rahmen dieser Arbeit erfolgt eine Klärung des Begriffs des stochastischen Denkens, eine anschauliche Darstellung der von Elart von Collani entwickelten Bernoullischen Stochastik und eine Beschreibung von Stochastikon. Dabei werden sowohl das Gesamtkonzept von Stochastikon, sowie die Ziele, Aufgaben und die Realisierung der beiden Teilsysteme namens Mentor und Encyclopedia vorgestellt. Das stochastische Denken erlaubt eine realitätsnahe Sichtweise der Dinge, d.h. eine Sichtweise, die mit den menschlichen Beobachtungen und Erfahrungen im Einklang steht und somit die Unsicherheit über zukünftige Entwicklungen berücksichtigt. Der in diesem Kontext verwendete Begriff der Unsicherheit bezieht sich ausschließlich auf zukünftige Entwicklungen und äußert sich in Variabilität. Quellen der Unsicherheit sind einerseits die menschliche Ignoranz und andererseits der Zufall. Unter Ignoranz wird hierbei die Unwissenheit des Menschen über die unbekannten, aber feststehenden Fakten verstanden, die die Anfangsbedingungen der zukünftigen Entwicklung repräsentieren. Die Bernoullische Stochastik liefert ein Regelwerk und ermöglicht die Entwicklung eines quantitativen Modells zur Beschreibung der Unsicherheit und expliziter Einbeziehung der beiden Quellen Ignoranz und Zufall. Das Modell trägt den Namen Bernoulli-Raum und bildet die Grundlage für die Herleitung quantitativer Verfahren, um zuverlässige und genaue Aussagen sowohl über die nicht-existente zufällige Zukunft (Vorhersageverfahren), als auch über die unbekannte feststehende Vergangenheit (Messverfahren). Das Softwaresystem Stochastikon implementiert die Bernoullische Stochastik in Form einer Reihe autarker, miteinander kommunizierender Teilsysteme. Ziel des Teilsystems Encyclopedia ist die Bereitstellung und Bewertung stochastischen Wissens. Das Teilsystem Mentor dient der Unterstützung des Anwenders bei der Problemlösungsfindung durch Identifikation eines richtigen Modells bzw. eines korrekten Bernoulli-Raums. Der Lösungsfindungsprozess selber enthält keinerlei Unsicherheit. Die ganze Unsicherheit steckt in der Lösung, d.h. im Bernoulli-Raum, der explizit die vorhandene Unwissenheit (Ignoranz) und den vorliegenden Zufall abdeckend enthält.
Overlay networks establish logical connections between users on top of the physical network. While randomly connected overlay networks provide only a best effort service, a new generation of structured overlay systems based on Distributed Hash Tables (DHTs) was proposed by the research community. However, there is still a lack of understanding the performance of such DHTs. Additionally, those architectures are highly distributed and therefore appear as a black box to the operator. Yet an operator does not want to lose control over his system and needs to be able to continuously observe and examine its current state at runtime. This work addresses both problems and shows how the solutions can be combined into a more self-organizing overlay concept. At first, we evaluate the performance of structured overlay networks under different aspects and thereby illuminate in how far such architectures are able to support carrier-grade applications. Secondly, to enable operators to monitor and understand their deployed system in more detail, we introduce both active as well as passive methods to gather information about the current state of the overlay network.
There is great interest in affordable, precise and reliable metrology underwater:
Archaeologists want to document artifacts in situ with high detail.
In marine research, biologists require the tools to monitor coral growth and geologists need recordings to model sediment transport.
Furthermore, for offshore construction projects, maintenance and inspection millimeter-accurate measurements of defects and offshore structures are essential.
While the process of digitizing individual objects and complete sites on land is well understood and standard methods, such as Structure from Motion or terrestrial laser scanning, are regularly applied, precise underwater surveying with high resolution is still a complex and difficult task.
Applying optical scanning techniques in water is challenging due to reduced visibility caused by turbidity and light absorption.
However, optical underwater scanners provide significant advantages in terms of achievable resolution and accuracy compared to acoustic systems.
This thesis proposes an underwater laser scanning system and the algorithms for creating dense and accurate 3D scans in water.
It is based on laser triangulation and the main optical components are an underwater camera and a cross-line laser projector.
The prototype is configured with a motorized yaw axis for capturing scans from a tripod.
Alternatively, it is mounted to a moving platform for mobile mapping.
The main focus lies on the refractive calibration of the underwater camera and laser projector, the image processing and 3D reconstruction.
For highest accuracy, the refraction at the individual media interfaces must be taken into account.
This is addressed by an optimization-based calibration framework using a physical-geometric camera model derived from an analytical formulation of a ray-tracing projection model.
In addition to scanning underwater structures, this work presents the 3D acquisition of semi-submerged structures and the correction of refraction effects.
As in-situ calibration in water is complex and time-consuming, the challenge of transferring an in-air scanner calibration to water without re-calibration is investigated, as well as self-calibration techniques for structured light.
The system was successfully deployed in various configurations for both static scanning and mobile mapping.
An evaluation of the calibration and 3D reconstruction using reference objects and a comparison of free-form surfaces in clear water demonstrate the high accuracy potential in the range of one millimeter to less than one centimeter, depending on the measurement distance.
Mobile underwater mapping and motion compensation based on visual-inertial odometry is demonstrated using a new optical underwater scanner based on fringe projection.
Continuous registration of individual scans allows the acquisition of 3D models from an underwater vehicle.
RGB images captured in parallel are used to create 3D point clouds of underwater scenes in full color.
3D maps are useful to the operator during the remote control of underwater vehicles and provide the building blocks to enable offshore inspection and surveying tasks.
The advancing automation of the measurement technology will allow non-experts to use it, significantly reduce acquisition time and increase accuracy, making underwater metrology more cost-effective.
Imagine a technology that automatically creates a full 3D thermal model of an environment and detects temperature peaks in it. For better orientation in the model it is enhanced with color information. The current state of the art for analyzing temperature related issues is thermal imaging. It is relevant for energy efficiency but also for securing important infrastructure such as power supplies and temperature regulation systems. Monitoring and analysis of the data for a large building is tedious as stable conditions need to be guaranteed for several hours and detailed notes about the pose and the environment conditions for each image must be taken. For some applications repeated measurements are necessary to monitor changes over time. The analysis of the scene is only possible through expertise and experience.
This thesis proposes a robotic system that creates a full 3D model of the environment with color and thermal information by combining thermal imaging with the technology of terrestrial laser scanning. The addition of a color camera facilitates the interpretation of the data and allows for other application areas. The data from all sensors collected at different positions is joined in one common reference frame using calibration and scan matching. The first part of the thesis deals with 3D point cloud processing with the emphasis on accessing point cloud data efficiently, detecting planar structures in the data and registering multiple point clouds into one common coordinate system. The second part covers the autonomous exploration and data acquisition with a mobile robot with the objective to minimize the unseen area in 3D space. Furthermore, the combination of different modalities, color images, thermal images and point cloud data through calibration is elaborated. The last part presents applications for the the collected data. Among these are methods to detect the structure of building interiors for reconstruction purposes and subsequent detection and classification of windows. A system to project the gathered thermal information back into the scene is presented as well as methods to improve the color information and to join separately acquired point clouds and photo series.
A full multi-modal 3D model contains all the relevant geometric information about the recorded scene and enables an expert to fully analyze it off-site. The technology clears the path for automatically detecting points of interest thereby helping the expert to analyze the heat flow as well as localize and identify heat leaks. The concept is modular and neither limited to achieving energy efficiency nor restricted to the use in combination with a mobile platform. It also finds its application in fields such as archaeology and geology and can be extended by further sensors.
Utilizing multiple access networks such as 5G, 4G, and Wi-Fi simultaneously can lead to increased robustness, resiliency, and capacity for mobile users. However, transparently implementing packet distribution over multiple paths within the core of the network faces multiple challenges including scalability to a large number of customers, low latency, and high-capacity packet processing requirements. In this paper, we offload congestion-aware multipath packet scheduling to a smartNIC. However, such hardware acceleration faces multiple challenges due to programming language and platform limitations. We implement different multipath schedulers in P4 with different complexity in order to cope with dynamically changing path capacities. Using testbed measurements, we show that our CMon scheduler, which monitors path congestion in the data plane and dynamically adjusts scheduling weights for the different paths based on path state information, can process more than 3.5 Mpps packets 25 μs latency.
A binary tanglegram is a drawing of a pair of rooted binary trees whose leaf sets are in one-to-one correspondence; matching leaves are connected by inter-tree edges. For applications, for example, in phylogenetics, it is essential that both trees are drawn without edge crossings and that the inter-tree edges have as few crossings as possible. It is known that finding a tanglegram with the minimum number of crossings is NP-hard and that the problem is fixed-parameter tractable with respect to that number.
We prove that under the Unique Games Conjecture there is no constant-factor approximation for binary trees. We show that the problem is NP-hard even if both trees are complete binary trees. For this case we give an O(n 3)-time 2-approximation and a new, simple fixed-parameter algorithm. We show that the maximization version of the dual problem for binary trees can be reduced to a version of MaxCut for which the algorithm of Goemans and Williamson yields a 0.878-approximation.
Historical maps are fascinating documents and a valuable source of information for scientists of various disciplines. Many of these maps are available as scanned bitmap images, but in order to make them searchable in useful ways, a structured representation of the contained information is desirable.
This book deals with the extraction of spatial information from historical maps. This cannot be expected to be solved fully automatically (since it involves difficult semantics), but is also too tedious to be done manually at scale.
The methodology used in this book combines the strengths of both computers and humans: it describes efficient algorithms to largely automate information extraction tasks and pairs these algorithms with smart user interactions to handle what is not understood by the algorithm. The effectiveness of this approach is shown for various kinds of spatial documents from the 16th to the early 20th century.
Small satellites contribute significantly in the rapidly evolving innovation in space engineering, in particular in distributed space systems for global Earth observation and communication services. Significant mass reduction by miniaturization, increased utilization of commercial high-tech components, and in particular standardization are the key drivers for modern miniature space technology.
This thesis addresses key fields in research and development on miniature satellite technology regarding efficiency, flexibility, and robustness. Here, these challenges are addressed by the University of Wuerzburg’s advanced pico-satellite bus, realizing a generic modular satellite architecture and standardized interfaces for all subsystems. The modular platform ensures reusability, scalability, and increased testability due to its flexible subsystem interface which allows efficient and compact integration of the entire satellite in a plug-and-play manner.
Beside systematic design for testability, a high degree of operational robustness is achieved by the consequent implementation of redundancy of crucial subsystems. This is combined with efficient fault detection, isolation and recovery mechanisms. Thus, the UWE-3 platform, and in particular the on-board data handling system and the electrical power system, offers one of the most efficient pico-satellite architectures launched in recent years and provides a solid basis for future extensions.
The in-orbit performance results of the pico-satellite UWE-3 are presented and summarize successful operations since its launch in 2013. Several software extensions and adaptations have been uploaded to UWE-3 increasing its capabilities. Thus, a very flexible platform for in-orbit software experiments and for evaluations of innovative concepts was provided and tested.
We use algebraic closures and structures which are derived from these in complexity theory. We classify problems with Boolean circuits and Boolean constraints according to their complexity. We transfer algebraic structures to structural complexity. We use the generation problem to classify important complexity classes.
Die Entwicklung eines wissensbasierten Systems, speziell eines Diagnosesystems, ist eine Teildisziplin der künstlichen Intelligenz und angewandten Informatik. Im Laufe der Forschung auf diesem Gebiet wurden verschiedene Lösungsansätze mit unterschiedlichem Erfolg bei der Anwendung in der Kraftfahrzeugdiagnose entwickelt. Diagnosesysteme in Vertragswerkstätten, das heißt in Fahrzeughersteller gebundenen Werkstätten, wenden hauptsächlich die fallbasierte Diagnostik an. Zum einen hält sich hier die Fahrzeugvielfalt in Grenzen und zum anderen besteht eine Meldepflicht bei neuen, nicht im System vorhandenen Fällen. Die freien Werkstätten verfügen nicht über eine solche Datenbank. Somit ist der fallbasierte Ansatz schwer umsetzbar. In freien Werkstätten - Fahrzeughersteller unabhängigen Werkstätten - basiert die Fehlersuche hauptsächlich auf Fehlerbäumen. Wegen der wachsenden Fahrzeugkomplexität, welche wesentlich durch die stark zunehmende Anzahl der durch mechatronische Systeme realisierten Funktionen bedingt ist, und der steigenden Typenvielfalt ist die geführte Fehlersuche in freien Werkstätten nicht immer zielführend. Um die Unterstützung des Personals von freien Werkstätten bei der zukünftigen Fehlersuche zu gewährleisten, werden neue Generationen von herstellerunabhängigen Diagnosetools benötigt, die die Probleme der Variantenvielfalt und Komplexität lösen. In der vorliegenden Arbeit wird ein Lösungsansatz vorgestellt, der einen qualitativen, modellbasierten Diagnoseansatz mit einem auf heuristischem Diagnosewissen basierenden Ansatz vereint. Neben der Grundlage zur Wissenserhebung werden in dieser Arbeit die theoretische Grundlage zur Beherrschung der Variantenvielfalt sowie die Tests für die erstellten Diagnosemodelle behandelt. Die Diagnose ist symptombasiert und die Inferenzmechanismen zur Verarbeitung des Diagnosewissens sind eine Kombination aus Propagierung der abweichenden physikalischen Größen im Modell und der Auswertung des heuristischen Wissens. Des Weiteren werden in dieser Arbeit verschiedene Aspekte der Realisierung der entwickelten theoretischen Grundlagen dargestellt, zum Beispiel: Systemarchitektur, Wissenserhebungsprozess, Ablauf des Diagnosevorgangs in den Werkstätten. Die Evaluierung der entwickelten Lösung bei der Wissenserhebung in Form von Modellerstellungen und Modellierungsworkshops sowie Feldtests dient nicht nur zur Bestätigung des entwickelten Ansatzes, sondern auch zur Ideenfindung für die Integration der entwickelten Tools in die existierende IT-Infrastruktur.
The rating of perceived exertion (RPE) is a subjective load marker and may assist in individualizing training prescription, particularly by adjusting running intensity. Unfortunately, RPE has shortcomings (e.g., underreporting) and cannot be monitored continuously and automatically throughout a training sessions. In this pilot study, we aimed to predict two classes of RPE (≤15 “Somewhat hard to hard” on Borg’s 6–20 scale vs. RPE >15 in runners by analyzing data recorded by a commercially-available smartwatch with machine learning algorithms. Twelve trained and untrained runners performed long-continuous runs at a constant self-selected pace to volitional exhaustion. Untrained runners reported their RPE each kilometer, whereas trained runners reported every five kilometers. The kinetics of heart rate, step cadence, and running velocity were recorded continuously ( 1 Hz ) with a commercially-available smartwatch (Polar V800). We trained different machine learning algorithms to estimate the two classes of RPE based on the time series sensor data derived from the smartwatch. Predictions were analyzed in different settings: accuracy overall and per runner type; i.e., accuracy for trained and untrained runners independently. We achieved top accuracies of 84.8 % for the whole dataset, 81.8 % for the trained runners, and 86.1 % for the untrained runners. We predict two classes of RPE with high accuracy using machine learning and smartwatch data. This approach might aid in individualizing training prescriptions.
Immersive virtual environments provide users with the opportunity to escape from the real world, but scripted dialogues can disrupt the presence within the world the user is trying to escape within. Both Non-Playable Character (NPC) to Player and NPC to NPC dialogue can be non-natural and the reliance on responding with pre-defined dialogue does not always meet the players emotional expectations or provide responses appropriate to the given context or world states. This paper investigates the application of Artificial Intelligence (AI) and Natural Language Processing to generate dynamic human-like responses within a themed virtual world. Each thematic has been analysed against humangenerated responses for the same seed and demonstrates invariance of rating across a range of model sizes, but shows an effect of theme and the size of the corpus used for fine-tuning the context for the game world.
This paper gives an overview of our recent activities in the field of satellite communication networks, including an introduction to geostationary satellite systems and Low Earth Orbit megaconstellations. To mitigate the high latencies of geostationary satellite networks, TCP-splitting Performance Enhancing Proxies are deployed. However, these cannot be applied in the case of encrypted transport headers as it is the case for VPNs or QUIC. We summarize performance evaluation results from multiple measurement campaigns. In a recently concluded project, multipath communication was used to combine the advantages of very heterogeneous communication paths: low data rate, low latency (e.g., DSL light) and high data rate, high latency (e.g., geostationary satellite).
The correct behavior of spacecraft components is the foundation of unhindered mission operation. However, no technical system is free of wear and degradation. A malfunction of one single component might significantly alter the behavior of the whole spacecraft and may even lead to a complete mission failure. Therefore, abnormal component behavior must be detected early in order to be able to perform counter measures. A dedicated fault detection system can be employed, as opposed to classical health monitoring, performed by human operators, to decrease the response time to a malfunction. In this paper, we present a generic model-based diagnosis system, which detects faults by analyzing the spacecraft’s housekeeping data. The observed behavior of the spacecraft components, given by the housekeeping data is compared to their expected behavior, obtained through simulation. Each discrepancy between the observed and the expected behavior of a component generates a so-called symptom. Given the symptoms, the diagnoses are derived by computing sets of components whose malfunction might cause the observed discrepancies. We demonstrate the applicability of the diagnosis system by using modified housekeeping data of the qualification model of an actual spacecraft and outline the advantages and drawbacks of our approach.
This work deals with teams in teleoperation scenarios, where one human team partner (supervisor) guides and controls multiple remote entities (either robotic or human) and coordinates their tasks. Such a team needs an appropriate infrastructure for sharing information and commands. The robots need to have a level of autonomy, which matches the assigned task. The humans in the team have to be provided with autonomous support, e.g. for information integration. Design and capabilities of the human-robot interfaces will strongly influence the performance of the team as well as the subjective feeling of the human team partners. Here, it is important to elaborate the information demand as well as how information is presented. Such human-robot systems need to allow the supervisor to gain an understanding of what is going on in the remote environment (situation awareness) by providing the necessary information. This includes achieving fast assessment of the robot´s or remote human´s state. Processing, integration and organization of data as well as suitable autonomous functions support decision making and task allocation and help to decrease the workload in this multi-entity teleoperation task. Interaction between humans and robots is improved by a common world model and a responsive system and robots. The remote human profits from a simplified user interface providing exactly the information needed for the actual task at hand. The topic of this thesis is the investigation of such teleoperation interfaces in human-robot teams, especially for high-risk, time-critical, and dangerous tasks. The aim is to provide a suitable human-robot team structure as well as analyze the demands on the user interfaces. On one side, it will be looked on the theoretical background (model, interactions, and information demand). On the other side, real implementations for system, robots, and user interfaces are presented and evaluated as testbeds for the claimed requirements. Rescue operations, more precisely fire-fighting, was chosen as an exemplary application scenario for this work. The challenges in such scenarios are high (highly dynamic environments, high risk, time criticality etc.) and it can be expected that results can be transferred to other applications, which have less strict requirements. The present work contributes to the introduction of human-robot teams in task-oriented scenarios, such as working in high risk domains, e.g. fire-fighting. It covers the theoretical background of the required system, the analysis of related human factors concepts, as well as discussions on implementation. An emphasis is placed on user interfaces, their design, requirements and user testing, as well as on the used techniques (three-dimensional sensor data representation, mixed reality, and user interface design guidelines). Further, the potential integration of 3D sensor data as well as the visualization on stereo visualization systems is introduced.
Mapping and localization of mobile robots in an unknown environment are essential for most high-level operations like autonomous navigation or exploration. This paper presents a novel approach for combining estimated trajectories, namely curvefusion. The robot used in the experiments is equipped with a horizontally mounted 2D profiler, a constantly spinning 3D laser scanner and a GPS module. The proposed algorithm first combines trajectories from different sensors to optimize poses of the planar three degrees of freedom (DoF) trajectory, which is then fed into continuous-time simultaneous localization and mapping (SLAM) to further improve the trajectory. While state-of-the-art multi-sensor fusion methods mainly focus on probabilistic methods, our approach instead adopts a deformation-based method to optimize poses. To this end, a similarity metric for curved shapes is introduced into the robotics community to fuse the estimated trajectories. Additionally, a shape-based point correspondence estimation method is applied to the multi-sensor time calibration. Experiments show that the proposed fusion method can achieve relatively better accuracy, even if the error of the trajectory before fusion is large, which demonstrates that our method can still maintain a certain degree of accuracy in an environment where typical pose estimation methods have poor performance. In addition, the proposed time-calibration method also achieves high accuracy in estimating point correspondences.
To jointly provide different services/technologies, like IP and Ethernet or IP and SDH/SONET, in a single network, equipment of multiple technologies needs to be deployed to the sites/Points of Presence (PoP) and interconnected with each other. Therein, a technology may provide transport functionality to other technologies and increase the number of available resources by using multiplexing techniques. By providing its own switching functionality, each technology creates connections in a logical layer which leads to the notion of multi-layer networks. The design of such networks comprises the deployment and interconnection of components to suit to given traffic demands. To prevent traffic loss due to failures of networking equipment, protection mechanisms need to be established. In multi-layer networks, protection usually can be applied in any of the considered layers. In turn, the hierarchical structure of multi-layer networks also bears shared risk groups (SRG). To achieve a cost-optimal resilient network, an appropriate combination of multiplexing techniques, technologies, and their interconnections needs to be found. Thus, network design is a combinatorial problem with a large parameter and solution space. After the design stage, the resources of a multi-layer network can be provided to traffic demands. Especially, dynamic capacity provisioning requires interaction of sites and layers, as well as accurate retrieval of constraint information. In recent years, generalized multiprotocol label switching (GMPLS) and path computation elements (PCE) have emerged as possible approaches for these challenges. Like the design, the provisioning of multi-layer networks comprises a variety of optimization parameters, like blocking probability, resilience, and energy efficiency. In this work, we introduce several efficient heuristics to approach the considered optimization problems. We perform capital expenditure (CAPEX)-aware design of multi-layer networks from scratch, based on IST NOBEL phase 2 project's cost and equipment data. We comprise traffic and resilience requirements in different and multiple layers as well as different network architectures. On top of the designed networks, we consider the dynamic provisioning of multi-layer traffic based on the GMPLS and PCE architecture. We evaluate different PCE deployments, information retrieval strategies, and re-optimization. Finally, we show how information about provisioning utilization can be used to provide a feedback for network design.
Dynamic point cloud compression based on projections, surface reconstruction and video compression
(2021)
In this paper we will present a new dynamic point cloud compression based on different projection types and bit depth, combined with the surface reconstruction algorithm and video compression for obtained geometry and texture maps. Texture maps have been compressed after creating Voronoi diagrams. Used video compression is specific for geometry (FFV1) and texture (H.265/HEVC). Decompressed point clouds are reconstructed using a Poisson surface reconstruction algorithm. Comparison with the original point clouds was performed using point-to-point and point-to-plane measures. Comprehensive experiments show better performance for some projection maps: cylindrical, Miller and Mercator projections.
Service orchestration requires enormous attention and is a struggle nowadays. Of course, virtualization provides a base level of abstraction for services to be deployable on a lot of infrastructures. With container virtualization, the trend to migrate applications to a micro-services level in order to be executable in Fog and Edge Computing environments increases manageability and maintenance efforts rapidly. Similarly, network virtualization adds effort to calibrate IP flows for Software-Defined Networks and eventually route it by means of Network Function Virtualization. Nevertheless, there are concepts like MAPE-K to support micro-service distribution in next-generation cloud and network environments. We want to explore, how a service distribution can be improved by adopting machine learning concepts for infrastructure or service changes. Therefore, we show how federated machine learning is integrated into a cloud-to-fog-continuum without burdening single nodes.
Mindfulness is considered an important factor of an individual's subjective well-being. Consequently, Human-Computer Interaction (HCI) has investigated approaches that strengthen mindfulness, i.e., by inventing multimedia technologies to support mindfulness meditation. These approaches often use smartphones, tablets, or consumer-grade desktop systems to allow everyday usage in users' private lives or in the scope of organized therapies. Virtual, Augmented, and Mixed Reality (VR, AR, MR; in short: XR) significantly extend the design space for such approaches. XR covers a wide range of potential sensory stimulation, perceptive and cognitive manipulations, content presentation, interaction, and agency. These facilities are linked to typical XR-specific perceptions that are conceptually closely related to mindfulness research, such as (virtual) presence and (virtual) embodiment. However, a successful exploitation of XR that strengthens mindfulness requires a systematic analysis of the potential interrelation and influencing mechanisms between XR technology, its properties, factors, and phenomena and existing models and theories of the construct of mindfulness. This article reports such a systematic analysis of XR-related research from HCI and life sciences to determine the extent to which existing research frameworks on HCI and mindfulness can be applied to XR technologies, the potential of XR technologies to support mindfulness, and open research gaps. Fifty papers of ACM Digital Library and National Institutes of Health's National Library of Medicine (PubMed) with and without empirical efficacy evaluation were included in our analysis. The results reveal that at the current time, empirical research on XR-based mindfulness support mainly focuses on therapy and therapeutic outcomes. Furthermore, most of the currently investigated XR-supported mindfulness interactions are limited to vocally guided meditations within nature-inspired virtual environments. While an analysis of empirical research on those systems did not reveal differences in mindfulness compared to non-mediated mindfulness practices, various design proposals illustrate that XR has the potential to provide interactive and body-based innovations for mindfulness practice. We propose a structured approach for future work to specify and further explore the potential of XR as mindfulness-support. The resulting framework provides design guidelines for XR-based mindfulness support based on the elements and psychological mechanisms of XR interactions.
In the last years, visual methods have been introduced in industrial software production and teaching of software engineering. In particular, the international standardization of a graphical software engineering language, the Unified Modeling Language (UML) was a reason for this tendency. Unfortunately, various problems exist in concrete realizations of tools, e.g. due to a missing compliance to the standard. One problem is the automatic layout, which is required for a consistent automatic software design. The thesis derives reasons and criteria for an automatic layout method, which produces drawings of UML class diagrams according to the UML specification and issues of human computer interaction, e.g. readability. A unique set of aesthetic criteria is combined from four different disciplines involved in this topic. Based on these aethetic rules, a hierarchical layout algorithm is developed, analyzed, measured by specialized measuring techniques and compared to related work. Then, the realization of the algorithm as a Java framework is given as an architectural description. Finally, adaptions to anticipated future changes of the UML, improvements of the framework and example drawings of the implementation are given.
Time-to-Live (TTL) caches decouple the occupancy of objects in cache through object-specific validity timers. Stateof- the art techniques provide exact methods for the calculation of object-specific hit probabilities given entire cache hierarchies with random inter-cache network delays. The system hit probability is a provider-centric metric as it relates to the origin offload, i.e., the decrease in the number of requests that are served by the content origin server. In this paper we consider a user-centric metric, i.e., the response time, which is shown to be structurally different from the system hit probability. Equipped with the state-of-theart exact modeling technique using Markov-arrival processes we derive expressions for the expected object response time and pave a way for its optimization under network delays.
Natural walking in virtual reality games is constrained by the physical boundaries defined by the size of the player’s tracking space. Impossible spaces, a redirected walking technique, enlarge the virtual environment by creating overlapping architecture and letting multiple locations occupy the same physical space. Within certain thresholds, this is subtle to the player. In this paper, we present our approach to implement such impossible spaces and describe how we handled challenges like objects with simulated physics or precomputed global illumination.
In der vorliegenden Arbeit wird das Problem der Kalibrierung Agenten-basierter Simulationen (ABS) behandelt, also das Problem, die Parameterwerte eines Agenten-basierten Simulationsmodells so einzustellen, dass valides Simulationsverhalten erreicht wird. Das Kalibrierungsproblem für Simulationen an sich ist nicht neu und ist im Rahmen klassischer Simulationsparadigmen, wie z.B. der Makro-Simulation, fester Bestandteil der Forschung. Im Vergleich zu den dort betrachteten Kalibrierungsproblemen zeichnet sich das Kalibrierungsproblem für ABS jedoch durch eine Reihe zusätzlicher Herausforderungen aus, welche die direkte Anwendung existierender Kalibrierungsverfahren in begrenzter Zeit erschweren, bzw. nicht mehr sinnvoll zulassen. Die Lösung dieser Probleme steht im Zentrum dieser Dissertation: Das Ziel besteht darin, den Nutzer bei der Kalibrierung von ABS auf der Basis von unzureichenden, potentiell fehlerhaften Daten und Wissen zu unterstützen. Dabei sollen drei Hauptprobleme gelöst werden: 1)Vereinfachung der Kalibrierung großer Agenten-Parametermengen auf der Mikro- Ebene in Agenten-basierten Simulationen durch Ausnutzung der spezifischen Struktur von ABS (nämlich dem Aufbau aus einer Menge von Agentenmodellen). 2)Kalibrierung Agenten-basierter Simulationen, so dass auf allen relevanten Beobachtungsebenen valides Simulationsverhalten erzeugt wird (mindestens Mikro und Makro-Ebene). Als erschwerende Randbedingung muss die Kalibrierung unter der Voraussetzung einer Makro-Mikro-Wissenslücke durchgeführt werden. 3)Kalibrierung Agenten-basierter Simulationen auf der Mikro-Ebene unter der Voraussetzung, dass zur Kalibrierung einzelner Agentenmodelle nicht ausreichend und potentiell verfälschte Daten zur Verhaltensvalidierung zur Verfügung stehen. Hierzu wird in dieser Arbeit das sogenannte Makro-Mikro-Verfahren zur Kalibrierung von Agenten-basierten Simulationen entwickelt. Das Verfahren besteht aus einem Basisverfahren, das im Verlauf der Arbeit um verschiedene Zusatzverfahren erweitert wird. Das Makro-Mikro-Verfahren und seine Erweiterungen sollen dazu dienen, die Modellkalibrierung trotz stark verrauschter Daten und eingeschränktem Wissen über die Wirkungszusammenhänge im Originalsystem geeignet zu ermöglichen und dabei den Kalibrierungsprozess zu beschleunigen: 1) Makro-Mikro-Kalibrierungsverfahren: Das in dieser Arbeit entwickelte Makro- Mikro-Verfahren unterstützt den Nutzer durch eine kombinierte Kalibrierung auf der Mikro- und der Makro-Beobachtungsebene, die gegebenenfalls durch Zwischenebenen erweitert werden kann. Der Grundgedanke des Verfahrens besteht darin, das Kalibrierungsproblem in eines auf aggregierter Verhaltensebene und eines auf der Ebene des Mikro-Agentenverhaltens aufzuteilen. Auf der Makro-Ebene wird nach validen idealen aggregierten Verhaltensmodellen (IVM) der Agenten gesucht. Auf der Mikro-Ebene wird versucht die individuellen Modelle der Agenten auf Basis des erwünschten Gesamtverhaltens und der ermittelten IVM so zu kalibrieren, das insgesamt Simulationsverhalten entsteht, das sowohl auf Mikro- als auch auf Makro-Ebene valide ist. 2) Erweiterung 1: Robuste Kalibrierung: Um den Umgang mit potentiell verrauschten Validierungskriterien (d.h. mit verrauschten Daten über ein Originalsystem, auf denen die Validierungskriterien der Simulation beruhen) und Modellteilen während der Kalibrierung von ABS zu ermöglichen, wird eine robuste Kalibrierungstechnik zur Anwendung im Makro-Mikro-Verfahren entwickelt. 3) Erweiterung 2: Kalibrierung mit Heterogenitätssuche: Als zweite Erweiterung des Makro-Mikro-Verfahrens wird ein Verfahren entwickelt, das das Problem des unklaren Detaillierungsgrades von ABS auf der Ebene der Parameterwerte adressiert. Prinzipiell kann zwar jeder Agent unterschiedliche Parameterwerte verwenden, obwohl eine geringere Heterogenität zur Erzeugung validen Verhaltens ausreichend wäre. Die entwickelte Erweiterung versucht, während der Kalibrierung, eine geeignete Heterogenitätsausprägung für die Parameterwerte der Agenten zu ermitteln. Unter einer Heterogenitätsausprägung wird dabei eine Einteilung der simulierten Agenten in Gruppen mit jeweils gleichen Parameterwerten verstanden. Die Heterogenitätssuche dient dazu, einen Kompromiss zu finden zwischen der Notwendigkeit, sehr große Parametersuchräume durchsuchen zu müssen und gleichzeitig den Suchraum so klein wie möglich halten zu wollen.
No abstract available
In many cases, problems, data, or information can be modeled as graphs. Graphs can be used as a tool for modeling in any case where connections between distinguishable objects occur. Any graph consists of a set of objects, called vertices, and a set of connections, called edges, such that any edge connects a pair of vertices. For example, a social network can be modeled by a graph by
transforming the users of the network into vertices and friendship relations between users into edges. Also physical networks like computer networks or transportation networks, for example, the metro network of a city, can be seen as graphs.
For making graphs and, thereby, the data that is modeled, well-understandable for users, we need a visualization. Graph drawing deals with algorithms for visualizing graphs. In this thesis, especially the use of crossings and curves is investigated for graph drawing problems under additional constraints. The constraints that occur in the problems investigated in this thesis especially restrict the positions of (a part of) the vertices; this is done either as a hard constraint or as an optimization criterion.
Digitization and transcription of historic documents offer new research opportunities for humanists and are the topics of many edition projects. However, manual work is still required for the main phases of layout recognition and the subsequent optical character recognition (OCR) of early printed documents. This paper describes and evaluates how deep learning approaches recognize text lines and can be extended to layout recognition using background knowledge. The evaluation was performed on five corpora of early prints from the 15th and 16th Centuries, representing a variety of layout features. While the main text with standard layouts could be recognized in the correct reading order with a precision and recall of up to 99.9%, also complex layouts were recognized at a rate as high as 90% by using background knowledge, the full potential of which was revealed if many pages of the same source were transcribed.
Given points in the plane, connect them using minimum ink. Though the task seems simple, it turns out to be very time consuming. In fact, scientists believe that computers cannot efficiently solve it. So, do we have to resign? This book examines such NP-hard network-design problems, from connectivity problems in graphs to polygonal drawing problems on the plane. First, we observe why it is so hard to optimally solve these problems. Then, we go over to attack them anyway. We develop fast algorithms that find approximate solutions that are very close to the optimal ones. Hence, connecting points with slightly more ink is not hard.
Knowledge-based systems (KBS) face an ever-increasing interest in various disciplines and contexts. Yet, the former aim to construct the ’perfect intelligent software’ continuously shifts to user-centered, participative solutions. Such systems enable users to contribute their personal knowledge to the problem solving process for increased efficiency and an ameliorated user experience. More precisely, we define non-functional key requirements of participative KBS as: Transparency (encompassing KBS status mediation), configurability (user adaptability, degree of user control/exploration), quality of the KB and UI, and evolvability (enabling the KBS to grow mature with their users). Many of those requirements depend on the respective target users, thus calling for a more user-centered development. Often, also highly expertise domains are targeted — inducing highly complex KBs — which requires a more careful and considerate UI/interaction design. Still, current KBS engineering (KBSE) approaches mostly focus on knowledge acquisition (KA) This often leads to non-optimal, little reusable, and non/little evaluated KBS front-end solutions.
In this thesis we propose a more encompassing KBSE approach. Due to the strong mutual influences between KB and UI, we suggest a novel form of intertwined UI and KB development. We base the approach on three core components for encompassing KBSE:
(1) Extensible prototyping, a tailored form of evolutionary prototyping; this builds on mature UI prototypes and offers two extension steps for the anytime creation of core KBS prototypes (KB + core UI) and fully productive KBS (core KBS prototype + common framing functionality). (2) KBS UI patterns, that define reusable solutions for the core KBS UI/interaction; we provide a basic collection of such patterns in this work. (3) Suitable usability instruments for the assessment of the KBS artifacts. Therewith, we do not strive for ’yet another’ self-contained KBS engineering methodology. Rather, we motivate to extend existing approaches by the proposed key components. We demonstrate this based on an agile KBSE model.
For practical support, we introduce the tailored KBSE tool ProKEt. ProKEt offers a basic selection of KBS core UI patterns and corresponding configuration options out of the box; their further adaption/extension is possible on various levels of expertise. For practical usability support, ProKEt offers facilities for quantitative and qualitative data collection. ProKEt explicitly fosters the suggested, intertwined development of UI and KB. For seamlessly integrating KA activities, it provides extension points for two selected external KA tools: For KnowOF, a standard office based KA environment. And for KnowWE, a semantic wiki for collaborative KA. Therewith, ProKEt offers powerful support for encompassing, user-centered KBSE.
Finally, based on the approach and the tool, we also developed a novel KBS type: Clarification KBS as a mashup of consultation and justification KBS modules. Those denote a specifically suitable realization for participative KBS in highly expertise contexts and consequently require a specific design. In this thesis, apart from more common UI solutions, we particularly also introduce KBS UI patterns especially tailored towards Clarification KBS.
How to Model and Predict the Scalability of a Hardware-In-The-Loop Test Bench for Data Re-Injection?
(2023)
This paper describes a novel application of an empirical network calculus model based on measurements of a hardware-in-the-loop (HIL) test system. The aim is to predict the performance of a HIL test bench for open-loop re-injection in the context of scalability. HIL test benches are distributed computer systems including software, hardware, and networking devices. They are used to validate complex technical systems, but have not yet been system under study themselves. Our approach is to use measurements from the HIL system to create an empirical model for arrival and service curves. We predict the performance and design the previously unknown parameters of the HIL simulator with network calculus (NC), namely the buffer sizes and the minimum needed pre-buffer time for the playback buffer. We furthermore show, that it is possible to estimate the CPU load from arrival and service-curves based on the utilization theorem, and hence estimate the scalability of the HIL system in the context of the number of sensor streams.
This paper presents a novel concept to extend state-of-the-art buffer monitoring with additional measures to estimate service-curves. The online algorithm for service-curve estimation replaces the state-of-the-art timestamp logging, as we expect it to overcome the main disadvantages of generating a huge amount of data and using a lot of CPU resources to store the data to a file during operation. We prove the accuracy of the online-algorithm offline with timestamp data and compare the derived bounds to the measured delay and backlog. We also do a proof-of- concept of the online-algorithm, implement it in LabVIEW and compare its performance to the timestamp logging by CPU load and data-size of the log-file. However, the implementation is still work-in-progress.
The present paper compares the effect of different waypoint parameters on the flight performance of a special autonomous indoor UAV (unmanned aerial vehicle) fusing ultrasonic, inertial, pressure and optical sensors for 3D positioning and controlling. The investigated parameters are the acceptance threshold for reaching a waypoint as well as the maximal waypoint step size or block size. The effect of these parameters on the flight time and accuracy of the flight path is investigated. Therefore the paper addresses how the acceptance threshold and step size influence the speed and accuracy of the autonomous flight and thus influence the performance of the presented autonomous quadrocopter under real indoor navigation circumstances.
Furthermore the paper demonstrates a drawback of the standard potential field method for navigation of such autonomous quadrocopters and points to an improvement.
A procedure to control all six DOF (degrees of freedom) of a UAV (unmanned aerial vehicle) without an external reference system and to enable fully autonomous flight is presented here. For 2D positioning the principle of optical flow is used. Together with the output of height estimation, fusing ultrasonic, infrared and inertial and pressure sensor data, the 3D position of the UAV can be computed, controlled and steered. All data processing is done on the UAV. An external computer with a pathway planning interface is for commanding purposes only. The presented system is part of the AQopterI8 project, which aims to develop an autonomous flying quadrocopter for indoor application. The focus of this paper is 2D positioning using an optical flow sensor. As a result of the performed evaluation, it can be concluded that for position hold, the standard deviation of the position error is 10cm and after landing the position error is about 30cm.