Refine
Has Fulltext
- yes (51)
Year of publication
- 2021 (51) (remove)
Document Type
- Journal article (28)
- Doctoral Thesis (13)
- Conference Proceeding (7)
- Book (1)
- Report (1)
- Working Paper (1)
Keywords
- virtual reality (9)
- artificial intelligence (3)
- augmented reality (3)
- Quality of Experience (2)
- Virtuelle Realität (2)
- avatars (2)
- human-computer interaction (2)
- immersion (2)
- immersive technologies (2)
- performance modeling (2)
- quality of experience (2)
- 3D Laser Scanning (1)
- 3D-reconstruction methods (1)
- 3DTK toolkit (1)
- Algorithmische Geometrie (1)
- Algorithmus (1)
- Angriff (1)
- Augmented Reality (1)
- Auto-Scaling (1)
- Außerschulische Bildung (1)
- Baseline Constrained LAMBDA (1)
- Benchmarking (1)
- Berechnungskomplexität (1)
- Beweissystem (1)
- Cloud Computing (1)
- Computer Science education (1)
- Computersicherheit (1)
- Computerspiel (1)
- Conjunction analysis (1)
- Corporate Network (1)
- CubeSat (1)
- CubeSat GNSS (1)
- Daedalus-Projekt (1)
- Data Mining (1)
- Datensicherung (1)
- Didaktik der Informatik (1)
- Dienstgüte (1)
- Disjoint pair (1)
- Drohne <Flugkörper> (1)
- EPM (1)
- Educational robotics (1)
- Educational robotics competitions (1)
- Eindringerkennung (1)
- Elementaggregation über dynamische Terme (1)
- Error-State Extendend Kalman Filter (1)
- Erweiterte Realität <Informatik> (1)
- FRAMEWORK <Programm> (1)
- Fachdidaktik (1)
- Feature Engineering & Extraction (1)
- Flugnavigation (1)
- Flugregelung (1)
- Forecasting (1)
- Game mechanic (1)
- Gamification (1)
- Graphenzeichnen (1)
- Gravitationsmodellunsicherheit (1)
- Gravity model uncertainty (1)
- HMD (Head-Mounted Display) (1)
- HTTP adaptive video streaming (1)
- IT-Sicherheit (1)
- Industrieroboter (1)
- Informatik (1)
- Integer Expression (1)
- Integer circuit (1)
- Intelligent Virtual Agents (1)
- Interkulturelles Lernen (1)
- International Comparative Research (1)
- Intrusion Detection (1)
- Kalman-Filter (1)
- Kerneldensity estimation (1)
- Knowledge encoding (1)
- Kombinatorik (1)
- Komplexität (1)
- Konjunktionsanalyse (1)
- Konvexe Zeichnungen (1)
- Künstliche Intelligenz (1)
- Lava (1)
- Lehrerbildung (1)
- Leistungsbewertung (1)
- Lernen (1)
- Loose Coupling (1)
- Lunar Caves (1)
- Lunar Exploration (1)
- Mapping (1)
- Markov model (1)
- Markovian and Non-Markovian systems (1)
- Mathematisches Modell (1)
- Medienkompetenz (1)
- Mensch-Maschine-Kommunikation (1)
- Mensch-Maschine-Schnittstelle (1)
- Modellgetriebene Entwicklung (1)
- Mond (1)
- NP-vollständiges Problem (1)
- Netzwerkdaten (1)
- Neuronale Netze (1)
- Nutzerschnittstellen (1)
- Onboard Software (1)
- Orakel <Informatik> (1)
- Orbit determination (1)
- Orbitbestimung (1)
- P-optimal (1)
- Phasenmehrdeutigkeit (1)
- Planare Graphen (1)
- Poisson surface reconstruction (1)
- Polyeder (1)
- Problemlösefähigkeiten (1)
- Prognose (1)
- Projektmanagement (1)
- Propositional proof system (1)
- Referenzmodell (1)
- Roboterwettbewerbe (1)
- Robotik (1)
- Rodos (1)
- SIMOC (1)
- Satellit (1)
- Sensorfusion (1)
- Serious game (1)
- Skalierbarkeit (1)
- Software-defined networking (1)
- Source Code Generation (1)
- Space Debris (1)
- Spherical Robot (1)
- Spielmechanik (1)
- Statische Analyse (1)
- TETCs (1)
- Telemetrie (1)
- Thermospheric density uncertainty (1)
- Thermosphärische Dichteunsicherheit (1)
- Ultra-Wideband (UWB) radio ranging (1)
- Umfrage (1)
- Uncertainty realism (1)
- Unmanned Aerial Vehicle (UAV) (1)
- Unsicherheitsrealismus (1)
- Veranstaltung (1)
- Videospiel (1)
- Videoübertragung (1)
- Virtualisierung (1)
- Vorgehensmodell (1)
- Wissensencodierung (1)
- XR (1)
- XR-artificial intelligence combination (1)
- XR-artificial intelligence continuum (1)
- Zeitreihenanalyse (1)
- acrophobia (1)
- affective computing (1)
- agency (1)
- agent-based models (1)
- agents (1)
- anxiety (1)
- application design (1)
- attack-aware (1)
- avatar embodiment (1)
- biosignals (1)
- clinical data warehouse (1)
- clinical study (1)
- co-authorships (1)
- co-inventorships (1)
- collaboration (1)
- communication network (1)
- communication networks (1)
- cost-sensitive learning (1)
- crowdsensing (1)
- crowdsourced measurements (1)
- decision-making (1)
- denial of service (1)
- dependable software (1)
- descriptors (1)
- design cycle (1)
- detection time simulation (1)
- dimensions of proximity (1)
- discrete-time models and analysis (1)
- education (1)
- educational games (1)
- electronic data capture (1)
- electronic health records (1)
- elevated plus-maze (1)
- embedding techniques (1)
- embodiment (1)
- emotions (1)
- event detection (1)
- foreign language learning and teaching (1)
- gait disorder (1)
- game mechanics (1)
- games (1)
- hackathons (1)
- handwriting (1)
- harness free satellite (1)
- head-mounted display (1)
- heterogeneous background (1)
- human behaviour (1)
- human computer interaction (HCI) (1)
- human-artificial intelligence interaction (1)
- human-artificial intelligence interface (1)
- human-centered design (1)
- human-centered, human-robot (1)
- imbalanced regression (1)
- immersive interfaces (1)
- immersive learning technologies (1)
- insect tracking (1)
- interactive authoring system (1)
- intercultural learning and teaching (1)
- interdisciplinary education (1)
- intrusion detection (1)
- iowa gambling task (1)
- laser ranging (1)
- learning environments (1)
- locomotion (1)
- machine learning (1)
- map projections (1)
- meditation (1)
- mindfulness (1)
- mixed reality (1)
- mixed-cultural settings (1)
- mobile networks (1)
- motivation (1)
- movement ecology (1)
- multimodal learning (1)
- multiple sclerosis (1)
- natural environment (1)
- natural language processing · · · (1)
- natural user interfaces (1)
- network function virtualization (1)
- octree (1)
- online survey (1)
- optimization (1)
- passive haptic feedback (1)
- performance analysis (1)
- performance evaluation (1)
- physiology (1)
- place-illusion (1)
- plausibility-illusion (1)
- point cloud compression (1)
- point cloud registration (1)
- point-to-plane measure (1)
- point-to-point measure (1)
- problem solving skills (1)
- procedural content generation (1)
- psychomotor training (1)
- psychophyisology (1)
- quality of experience prediction (1)
- queueing theory (1)
- real-time (1)
- realism (1)
- recommender system (1)
- rehabilitation (1)
- research methods (1)
- sample weighting (1)
- satellite technology (1)
- secondary data usage (1)
- self-assembly (1)
- self-aware (1)
- self-supervised learning (1)
- semantic understanding (1)
- sensor devices (1)
- serious games (1)
- sesnsors (1)
- sketching (1)
- software-definded networking (1)
- space–terrestrial networks (1)
- spatial presence (1)
- statistical methods (1)
- statistical validity (1)
- stochastic processes (1)
- stroke (1)
- structural battery (1)
- stylus (1)
- supervised learning (1)
- system simulation (1)
- systematic literature review (1)
- systematic review (1)
- teacher education (1)
- time perception (1)
- tools (1)
- trait anxiety (1)
- translational neuroscience (1)
- transportation (1)
- usability evaluation (1)
- use cases (1)
- user experience (1)
- user interaction (1)
- user study (1)
- user-generated content (1)
- verbal behaviour (1)
- video game QoE (1)
- video game context factors (1)
- video streaming (1)
- virtual body ownership (1)
- virtual environments (1)
- virtual humans (1)
- virtual-reality-continuum (1)
- wireless communication (1)
- wireless-bus (1)
Institute
- Institut für Informatik (51) (remove)
Sonstige beteiligte Institutionen
EU-Project number / Contract (GA) number
- 824128 (1)
Das Management von Projekten, welche sowohl einmalige und interdisziplinäre Aufgabenstellungen als auch individuelle Rahmenbedingungen und Einschränkungen umfassen, stellt eine anspruchsvolle Aufgabe dar. Es gibt einige standardisierte Vorgehensmodelle, die einen organisatorischen Rahmen aus Phasen, Prozessen, Rollen und anzuwendenden Methoden anbieten.
Traditionellen Vorgehensmodellen wird in der Regel gefolgt, wenn die zu erzielenden Ergebnisse und der Ablauf eines Projektes auf Basis der zur Verfügung stehenden Informationen geplant werden können.
Agile Vorgehensmodelle werden vorranging genutzt, wenn keine ausreichenden Informationen zur Verfügung stehen, um eine vollständige Planung aufzusetzen. Ihr Fokus liegt darauf, flexibel auf sich ändernde Anforderungen einzugehen. Im direkten Austausch mit Kunden werden in meist mehreren aufeinander folgenden Zyklen Zwischenergebnisse bewertet und darauf basierend die jeweils nächsten Entwicklungsschritte geplant und umgesetzt.
Hybride Vorgehensmodelle werden genutzt, wenn Methoden aus mehreren unterschiedlichen Vorgehensmodellen erforderlich sind, um ein Projekt zu bearbeiten.
Die Bedeutung hybrider Vorgehensmodelle hat über die Jahre immer weiter zugenommen. Ihr besonderer Nutzen liegt darin, dass die Methodenauswahl auf den individuellen Kontext eines Projektes angepasst werden kann. Da es in der Praxis aber eine sehr große Anzahl an Methoden gibt, ist die Auswahl der zum Kontext passenden und deren Kombination zu einem individuellen Vorgehensmodell selbst für Experten/-innen eine Herausforderung. Die Forschungsergebnisse der vorliegenden Arbeit zeigen, dass es bisher auch kein Schema zur Unterstützung dieses Prozesses gab.
Um diese Forschungslücke zu schließen, wurde ein adaptives Referenzmodell für hybrides Projektmanagement (ARHP) entwickelt. Der wissenschaftliche Beitrag besteht zum einen in der Entwicklung eines Ablaufs zur Selektion und Kombination von zum Kontext passenden Methoden und zum anderen in der Umsetzung des Ablaufs als semi-automatisches Werkzeug. Referenzmodellnutzer/-innen können darin ihren individuellen Projektkontext durch die Auswahl zutreffender Kriterien (sogenannter Parameterausprägungen) erfassen. Das ARHP bietet ihnen dann ein Vorgehensmodell an, welches aus miteinander anwendbaren und verknüpfbaren Methoden besteht.
Da in der Projektmanagement Community häufig schnelle Entscheidungen für ein geeignetes Vorgehensmodell erforderlich sind und selbst Experten/-innen nicht alle Methoden kennen, wird der Nutzen der ''digitalen Beratung'', die das semi-automatische ARHP bietet, als hoch eingestuft.
Sowohl die für die Erfassung des Kontextes erforderlichen Parameter als auch die Methoden mit der höchsten Praxisrelevanz, wurden anhand einer umfangreichen Umfrage erforscht. Ihr wissenschaftlicher Beitrag besteht unter anderem in der erstmaligen Erfassung von Begründungen für die Verwendung von Methoden im Rahmen individueller, hybrider Vorgehensmodelle. Zudem erlauben die gesammelten Daten einen direkten Vergleich der Methodennutzung in funktionierenden und nicht funktionierenden hybriden Vorgehensmodellen.
Mit der so vorhandenen Datengrundlage wird in drei Design Science Research Zyklen ein Algorithmus entwickelt, der den Adaptionsmechanismus des ARHP bildet. Die Evaluation des ARHP erfolgt anhand des entwickelten semi-automatischen Prototypen unter Einbeziehung von Projektmanagementexperten/-innen.
Ausführungen zur Pflege des ARHP können als Handlungsanleitung für Referenzmodellkonstrukteure/-innen verstanden werden. Sie bilden den letzten Teil der Arbeit und zeigen, wie das ARHP kontinuierlich weiterentwickelt werden kann. Zudem wird ein Ausblick darauf gegeben, um welche Themen das ARHP im Rahmen weiterführender Forschung erweitert werden kann. Dabei handelt es sich zum Beispiel um eine noch stärkere Automatisierung und Empfehlungen für das Change Management, welche beide bereits in Vorbereitung sind.
Produktionssysteme mit Industrierobotern werden zunehmend komplex; waren deren Arbeitsbereiche früher noch statisch und abgeschirmt, und die programmierten Abläufe gleichbleibend, so sind die Anforderungen an moderne Robotik-Produktionsanlagen gestiegen: Diese sollen sich jetzt mithilfe von intelligenter Sensorik auch in unstrukturierten Umgebungen einsetzen lassen, sich bei sinkenden Losgrößen aufgrund individualisierter Produkte und häufig ändernden Produktionsaufgaben leicht rekonfigurieren lassen, und sogar eine direkte Zusammenarbeit zwischen Mensch und Roboter ermöglichen. Gerade auch bei dieser Mensch-Roboter-Kollaboration wird es damit notwendig, dass der Mensch die Daten und Aktionen des Roboters leicht verstehen kann. Aufgrund der gestiegenen Anforderungen müssen somit auch die Bedienerschnittstellen dieser Systeme verbessert werden. Als Grundlage für diese neuen Benutzerschnittstellen bietet sich Augmented Reality (AR) als eine Technologie an, mit der sich komplexe räumliche Daten für den Bediener leicht verständlich darstellen lassen. Komplexe Informationen werden dabei in der Arbeitsumgebung der Nutzer visualisiert und als virtuelle Einblendungen sichtbar gemacht, und so auf einen Blick verständlich. Die diversen existierenden AR-Anzeigetechniken sind für verschiedene Anwendungsfelder unterschiedlich gut geeignet, und sollten daher flexibel kombinier- und einsetzbar sein. Auch sollen diese AR-Systeme schnell und einfach auf verschiedenartiger Hardware in den unterschiedlichen Arbeitsumgebungen in Betrieb genommen werden können. In dieser Arbeit wird ein Framework für Augmented Reality Systeme vorgestellt, mit dem sich die genannten Anforderungen umsetzen lassen, ohne dass dafür spezialisierte AR-Hardware notwendig wird. Das Flexible AR-Framework kombiniert und bündelt dafür verschiedene Softwarefunktionen für die grundlegenden AR-Anzeigeberechnungen, für die Kalibrierung der notwendigen Hardware, Algorithmen zur Umgebungserfassung mittels Structured Light sowie generische ARVisualisierungen und erlaubt es dadurch, verschiedene AR-Anzeigesysteme schnell und flexibel in Betrieb zu nehmen und parallel zu betreiben. Im ersten Teil der Arbeit werden Standard-Hardware für verschiedene AR-Visualisierungsformen sowie die notwendigen Algorithmen vorgestellt, um diese flexibel zu einem AR-System zu kombinieren. Dabei müssen die einzelnen verwendeten Geräte präzise kalibriert werden; hierfür werden verschiedene Möglichkeiten vorgestellt, und die mit ihnen dann erreichbaren typischen Anzeige- Genauigkeiten in einer Evaluation charakterisiert. Nach der Vorstellung der grundlegenden ARSysteme des Flexiblen AR-Frameworks wird dann eine Reihe von Anwendungen vorgestellt, bei denen das entwickelte System in konkreten Praxis-Realisierungen als AR-Benutzerschnittstelle zum Einsatz kam, unter anderem zur Überwachung von, Zusammenarbeit mit und einfachen Programmierung von Industrierobotern, aber auch zur Visualisierung von komplexen Sensordaten oder zur Fernwartung. Im Verlauf der Arbeit werden dadurch die Vorteile, die sich durch Verwendung der AR-Technologie in komplexen Produktionssystemen ergeben, herausgearbeitet und in Nutzerstudien belegt.
Uplink vs. Downlink: Machine Learning-Based Quality Prediction for HTTP Adaptive Video Streaming
(2021)
Streaming video is responsible for the bulk of Internet traffic these days. For this reason, Internet providers and network operators try to make predictions and assessments about the streaming quality for an end user. Current monitoring solutions are based on a variety of different machine learning approaches. The challenge for providers and operators nowadays is that existing approaches require large amounts of data. In this work, the most relevant quality of experience metrics, i.e., the initial playback delay, the video streaming quality, video quality changes, and video rebuffering events, are examined using a voluminous data set of more than 13,000 YouTube video streaming runs that were collected with the native YouTube mobile app. Three Machine Learning models are developed and compared to estimate playback behavior based on uplink request information. The main focus has been on developing a lightweight approach using as few features and as little data as possible, while maintaining state-of-the-art performance.
Constraining graph layouts - that is, restricting the placement of vertices and the routing of edges to obey certain constraints - is common practice in graph drawing.
In this book, we discuss algorithmic results on two different restriction types:
placing vertices on the outer face and on the integer grid.
For the first type, we look into the outer k-planar and outer k-quasi-planar graphs, as well as giving a linear-time algorithm to recognize full and closed outer k-planar graphs Monadic Second-order Logic.
For the second type, we consider the problem of transferring a given planar drawing onto the integer grid while perserving the original drawings topology;
we also generalize a variant of Cauchy's rigidity theorem for orthogonal polyhedra of genus 0 to those of arbitrary genus.
Effects of Acrophobic Fear and Trait Anxiety on Human Behavior in a Virtual Elevated Plus-Maze
(2021)
The Elevated Plus-Maze (EPM) is a well-established apparatus to measure anxiety in rodents, i.e., animals exhibiting an increased relative time spent in the closed vs. the open arms are considered anxious. To examine whether such anxiety-modulated behaviors are conserved in humans, we re-translated this paradigm to a human setting using virtual reality in a Cave Automatic Virtual Environment (CAVE) system. In two studies, we examined whether the EPM exploration behavior of humans is modulated by their trait anxiety and also assessed the individuals’ levels of acrophobia (fear of height), claustrophobia (fear of confined spaces), sensation seeking, and the reported anxiety when on the maze. First, we constructed an exact virtual copy of the animal EPM adjusted to human proportions. In analogy to animal EPM studies, participants (N = 30) freely explored the EPM for 5 min. In the second study (N = 61), we redesigned the EPM to make it more human-adapted and to differentiate influences of trait anxiety and acrophobia by introducing various floor textures and lower walls of closed arms to the height of standard handrails. In the first experiment, hierarchical regression analyses of exploration behavior revealed the expected association between open arm avoidance and Trait Anxiety, an even stronger association with acrophobic fear. In the second study, results revealed that acrophobia was associated with avoidance of open arms with mesh-floor texture, whereas for trait anxiety, claustrophobia, and sensation seeking, no effect was detected. Also, subjects’ fear rating was moderated by all psychometrics but trait anxiety. In sum, both studies consistently indicate that humans show no general open arm avoidance analogous to rodents and that human EPM behavior is modulated strongest by acrophobic fear, whereas trait anxiety plays a subordinate role. Thus, we conclude that the criteria for cross-species validity are met insufficiently in this case. Despite the exploratory nature, our studies provide in-depth insights into human exploration behavior on the virtual EPM.
The recently published ITU-T Recommendation G1.032 proposes a list of factors that may influence cloud and online gaming Quality of Experience (QoE). This paper provides two practical evaluations of proposed system and context influence factors: First, it investigates through an online survey (n=488) the popularity of platforms, preferred ways of distribution, and motivational aspects including subjective valuations of characteristics offered by today's prevalent gaming platforms. Second, the paper evaluates a large dataset of objective metrics for various gaming platforms: game lists, playthrough lengths, prices, etc., and contrasts these metrics against the gamers' opinions. The combined data-driven approach presented in this paper complements in-person and lab studies usually employed.
In the past two decades, there has been a trend to move from traditional television to Internet-based video services. With video streaming becoming one of the most popular applications in the Internet and the current state of the art in media consumption, quality expectations of consumers are increasing. Low quality videos are no longer considered acceptable in contrast to some years ago due to the increased sizes and resolution of devices. If the high expectations of the users are not met and a video is delivered in poor quality, they often abandon the service. Therefore, Internet Service Providers (ISPs) and video service providers are facing the challenge of providing seamless multimedia delivery in high quality. Currently, during peak hours, video streaming causes almost 58\% of the downstream traffic on the Internet. With higher mobile bandwidth, mobile video streaming has also become commonplace. According to the 2019 Cisco Visual Networking Index, in 2022 79% of mobile traffic will be video traffic and, according to Ericsson, by 2025 video is forecasted to make up 76% of total Internet traffic. Ericsson further predicts that in 2024 over 1.4 billion devices will be subscribed to 5G, which will offer a downlink data rate of 100 Mbit/s in dense urban environments.
One of the most important goals of ISPs and video service providers is for their users to have a high Quality of Experience (QoE). The QoE describes the degree of delight or annoyance a user experiences when using a service or application. In video streaming the QoE depends on how seamless a video is played and whether there are stalling events or quality degradations. These characteristics of a transmitted video are described as the application layer Quality of Service (QoS). In general, the QoS is defined as "the totality of characteristics of a telecommunications service that bear on its ability to satisfy stated and implied needs of the user of the service" by the ITU. The network layer QoS describes the performance of the network and is decisive for the application layer QoS.
In Internet video, typically a buffer is used to store downloaded video segments to compensate for network fluctuations. If the buffer runs empty, stalling occurs. If the available bandwidth decreases temporarily, the video can still be played out from the buffer without interruption. There are different policies and parameters that determine how large the buffer is, at what buffer level to start the video, and at what buffer level to resume playout after stalling. These have to be finely tuned to achieve the highest QoE for the user. If the bandwidth decreases for a longer time period, a limited buffer will deplete and stalling can not be avoided. An important research question is how to configure the buffer optimally for different users and situations. In this work, we tackle this question using analytic models and measurement studies. With HTTP Adaptive Streaming (HAS), the video players have the capability to adapt the video bit rate at the client side according to the available network capacity. This way the depletion of the video buffer and thus stalling can be avoided. In HAS, the quality in which the video is played and the number of quality switches also has an impact on the QoE. Thus, an important problem is the adaptation of video streaming so that these parameters are optimized. In a shared WiFi multiple video users share a single bottleneck link and compete for bandwidth. In such a scenario, it is important that resources are allocated to users in a way that all can have a similar QoE. In this work, we therefore investigate the possible fairness gain when moving from network fairness towards application-layer QoS fairness. In mobile scenarios, the energy and data consumption of the user device are limited resources and they must be managed besides the QoE. Therefore, it is also necessary, to investigate solutions, that conserve these resources in mobile devices. But how can resources be conserved without sacrificing application layer QoS? As an example for such a solution, this work presents a new probabilistic adaptation algorithm that uses abandonment statistics for ts decision making, aiming at minimizing the resource consumption while maintaining high QoS.
With current protocol developments such as 5G, bandwidths are increasing, latencies are decreasing and networks are becoming more stable, leading to higher QoS. This allows for new real time data intensive applications such as cloud gaming, virtual reality and augmented reality applications to become feasible on mobile devices which pose completely new research questions. The high energy consumption of such applications still remains an issue as the energy capacity of devices is currently not increasing as quickly as the available data rates. In this work we compare the optimal performance of different strategies for adaptive 360-degree video streaming.
Computer games are highly immersive, engaging, and motivating learning environments. By providing a tutorial at the start of a new game, players learn the basics of the game's underlying principles as well as practice how to successfully play the game. During the actual gameplay, players repetitively apply this knowledge, thus improving it due to repetition. Computer games also challenge players with a constant stream of new challenges which increase in difficulty over time. As a result, computer games even require players to transfer their knowledge to master these new challenges. A computer game consists of several game mechanics. Game mechanics are the rules of a computer game and encode the game's underlying principles. They create the virtual environments, generate a game's challenges and allow players to interact with the game. Game mechanics also can encode real world knowledge. This knowledge may be acquired by players via gameplay. However, the actual process of knowledge encoding and knowledge learning using game mechanics has not been thoroughly defined, yet. This thesis therefore proposes a theoretical model to define the knowledge learning using game mechanics: the Gamified Knowledge Encoding. The model is applied to design a serious game for affine transformations, i.e., GEtiT, and to predict the learning outcome of playing a computer game that encodes orbital mechanics in its game mechanics, i.e., Kerbal Space Program. To assess the effects of different visualization technologies on the overall learning outcome, GEtiT visualizes the gameplay in desktop-3D and immersive virtual reality. The model's applicability for effective game design as well as GEtiT's overall design are evaluated in a usability study. The learning outcome of playing GEtiT and Kerbal Space Program is assessed in four additional user studies. The studies' results validate the use of the Gamified Knowledge Encoding for the purpose of developing effective serious games and to predict the learning outcome of existing serious games. GEtiT and Kerbal Space Program yield a similar training effect but a higher motivation to tackle the assignments in comparison to a traditional learning method. In conclusion, this thesis expands the understanding of using game mechanics for an effective learning of knowledge. The presented results are of high importance for researches, educators, and developers as they also provide guidelines for the development of effective serious games.
The successful development and classroom integration of Virtual (VR) and Augmented Reality (AR) learning environments requires competencies and content knowledge with respect to media didactics and the respective technologies. The paper discusses a pedagogical concept specifically aiming at the interdisciplinary education of pre-service teachers in collaboration with human-computer interaction students. The students’ overarching goal is the interdisciplinary realization and integration of VR/AR learning environments in teaching and learning concepts. To assist this approach, we developed a specific tutorial guiding the developmental process. We evaluate and validate the effectiveness of the overall pedagogical concept by analyzing the change in attitudes regarding 1) the use of VR/AR for educational purposes and in competencies and content knowledge regarding 2) media didactics and 3) technology. Our results indicate a significant improvement in the knowledge of media didactics and technology. We further report on four STEM learning environments that have been developed during the seminar.
Impaired decision-making leads to the inability to distinguish between advantageous and disadvantageous choices. The impairment of a person’s decision-making is a common goal of gambling games. Given the recent trend of gambling using immersive Virtual Reality it is crucial to investigate the effects of both immersion and the virtual environment (VE) on decision-making. In a novel user study, we measured decision-making using three virtual versions of the Iowa Gambling Task (IGT). The versions differed with regard to the degree of immersion and design of the virtual environment. While emotions affect decision-making, we further measured the positive and negative affect of participants. A higher visual angle on a stimulus leads to an increased emotional response. Thus, we kept the visual angle on the Iowa Gambling Task the same between our conditions. Our results revealed no significant impact of immersion or the VE on the IGT. We further found no significant difference between the conditions with regard to positive and negative affect. This suggests that neither the medium used nor the design of the VE causes an impairment of decision-making. However, in combination with a recent study, we provide first evidence that a higher visual angle on the IGT leads to an effect of impairment.
Having a mixed-cultural membership becomes increasingly common in our modern society. It is thus beneficial in several ways to create Intelligent Virtual Agents (IVAs) that reflect a mixed-cultural background as well, e.g., for educational settings. For research with such IVAs, it is essential that they are classified as non-native by members of a target culture. In this paper, we focus on variations of IVAs’ speech to create the impression of non-native speakers that are identified as such by speakers of two different mother tongues. In particular, we investigate grammatical mistakes and identify thresholds beyond which the agents is clearly categorised as a non-native speaker. Therefore, we conducted two experiments: one for native speakers of German, and one for native speakers of English. Results of the German study indicate that beyond 10% of word order mistakes and 25% of infinitive mistakes German-speaking IVAs are perceived as non-native speakers. Results of the English study indicate that beyond 50% of omission mistakes and 50% of infinitive mistakes English-speaking IVAs are perceived as non-native speakers. We believe these thresholds constitute helpful guidelines for computational approaches of non-native speaker generation, simplifying research with IVAs in mixed-cultural settings.
Educational robotics is an innovative approach to teaching and learning a variety of different concepts and skills as well as motivating students in the field of Science, Technology, Engineering, and Mathematics (STEM) education. This especially applies to educational robotics competitions such as, for example, the FIRST LEGO League, the RoboCup Junior, or the World Robot Olympiad as out-of-school and goal-oriented approach to educational robotics. These competitions have gained greatly in popularity in recent years and thousands of students participate in these competitions worldwide each year. Moreover, the corresponding technology became more accessible for teachers and students to use it in their classrooms and has arguably a high potential to impact the nature of science education at all levels. One skill, which is said to be benefitting from educational robotics, is problem solving. This thesis understands problem solving skills as engineering design skills (in contrast to scientific inquiry). Problem solving skills count as important skills as demanded by industry leaders and policy makers in the context of 21st century skills, which are relevant for students to be well-prepared for their future working life in today’s world, shaped by an ongoing process of automation, globalization, and digitalization. The overall aim of this thesis is to try to answer the question if educational robotics competitions such as the World Robot Olympiad (WRO) have a positive impact on students’ learning in terms of their problem solving skills (as part of 21st century skills). In detail, this thesis focuses on a) if students can improve their problem solving skills through participation in educational robotics competitions, b) how this skill development is accomplished, and c) the teachers’ support of their students during their learning process in the competition. The corresponding empirical studies were conducted throughout the seasons of 2018 and 2019 of the WRO in Germany. The results show overall positive effects of the participation in the WRO on students’ learning of problem solving skills. They display an increase of students’ problem solving skills, which is not moderated by other variables such as the competition’s category or age group, the students’ gender or experience, or the success of the teams at the competition. Moreover, the results indicate that students develop their problem solving skills by using a systematic engineering design process and sophisticated problem solving strategies. Lastly, the teacher’s role in the educational robotics competitions as manager and guide (in terms of the constructionist learning theory) of the students’ learning process (especially regarding the affective level) is underlined by the results of this thesis. All in all, this thesis contributes to the research gap concerning the lack of systematic evaluation of educational robotics to promote students’ learning by providing more (methodologically) sophisticated research on this topic. Thereby, this thesis follows the call for more rigorous (quantitative) research by the educational robotics community, which is necessary to validate the impact of educational robotics.
The presence of a partner can attenuate physiological fear responses, a phenomenon known as social buffering. However, not all individuals are equally sociable. Here we investigated whether social buffering of fear is shaped by sensitivity to social anxiety (social concern) and whether these effects are different in females and males. We collected skin conductance responses (SCRs) and affect ratings of female and male participants when they experienced aversive and neutral sounds alone (alone treatment) or in the presence of an unknown person of the same gender (social treatment). Individual differences in social concern were assessed based on a well-established questionnaire. Our results showed that social concern had a stronger effect on social buffering in females than in males. The lower females scored on social concern, the stronger the SCRs reduction in the social compared to the alone treatment. The effect of social concern on social buffering of fear in females disappeared if participants were paired with a virtual agent instead of a real person. Together, these results showed that social buffering of human fear is shaped by gender and social concern. In females, the presence of virtual agents can buffer fear, irrespective of individual differences in social concern. These findings specify factors that shape the social modulation of human fear, and thus might be relevant for the treatment of anxiety disorders.
E-Mails, Online Banking und Videokonferenzen sind aus unserem heutigen Alltag nicht mehr wegzudenken. Bei all diesen Aktivitäten werden zahlreiche personenbezogene Informationen und vertrauenswürdige Daten digital übertragen und gespeichert. Zur Sicherstellung der digitalen Daten vor unbefugten Zugriffen und Manipulationen existieren verschiedenste Konzepte, Methoden und Verfahren, die sich unter dem Begriff IT-Sicherheit zusammenfassen lassen. Klassische Sicherheitslösungen aus dem Bereich IT-Sicherheit sind Firewalls und Virenscanner. Derartige Ansätze sind meist regelbasiert und prüfen Dateien beziehungsweise eingehenden Netzwerkverkehr anhand einer Liste bekannter Angriffssignaturen. Folglich können diese Systeme nur bereits bekannte Angriffsszenarien detektieren und bieten keinen Schutz vor neuartigen Angriffen. Somit entsteht im Bereich IT-Sicherheit ein Wettlauf zwischen Hackern und IT-Sicherheitsexperten, bei dem die Hacker stets nach neuen Mitteln und Wegen suchen, die existierenden Sicherheitslösungen zu überwinden, während IT-Sicherheitsexperten stetig ihre Schutzmechanismen verbessern.
Die vorliegende Arbeit widmet sich der Detektion von Angriffsszenarien in Unternehmensnetzwerken mithilfe von Data Mining-Methoden. Diese Methoden sind in der Lage anhand von repräsentativen Daten die darin enthaltenen Strukturen zu erlernen und zu generalisieren. Folglich können sich Data Mining-Methoden grundsätzlich zur Detektion neuer Angriffsszenarien eignen, wenn diese Angriffsszenarien Überschneidungen mit bekannten Angriffsszenarien aufweisen oder sich wesentlich vom bekannten Normalverhalten unterscheiden. In dieser Arbeit werden netzwerkbasierte Daten im NetFlow Format analysiert, da diese einen aggregierten Überblick über das Geschehen im Netzwerk bieten. Häufig können Netzwerkdaten aufgrund datenschutzrechtlicher Bedenken nicht veröffentlicht werden, was für die Erzeugung synthetischer, aber realistischer Netzwerkdaten spricht. Des Weiteren führt die Beschaffenheit der Netzwerkdaten dazu, dass eine Kombination von kontinuierlichen und kategorischen Attributen analysiert werden muss, was vor allem das Vergleichen der Daten bezüglich ihrer Ähnlichkeit erschwert.
Diese Arbeit liefert methodische Beiträge zu jeder der drei genannten Herausforderungen. Im Bereich der Abstandsberechnung kategorischer Werte werden mit ConDist und IP2Vec zwei unterschiedliche Ansätze entwickelt. ConDist ist ein universell einsetzbares Abstandsmaß zur Berechnung von Abständen zwischen Datenpunkten, die aus kontinuierlichen und kategorischen Attributen bestehen. IP2Vec ist auf Netzwerkdaten spezialisiert und transformiert kategorische Werte in kontinuierliche Vektoren.
Im Bereich der Generierung realistischer Netzwerkdaten werden neben einer ausführlichen Literaturrecherche zwei unterschiedliche Ansätze vorgestellt. Zunächst wird ein auf Simulation basierter Ansatz zur Generierung flowbasierter Datensätze entwickelt. Dieser Ansatz basiert auf einer Testumgebung und simuliert typische Benutzeraktivitäten durch automatisierte Python Skripte. Parallel hierzu wird ein zweiter Ansatz zur synthetischen Generierung flowbasierter Netzwerkdaten durch Modellierung mithilfe von Generative Adversarial Networks entwickelt. Dieser Ansatz erlernt die zugrundeliegenden Eigenschaften der Netzwerkdaten und ist anschließend in der Lage, neue Netzwerkdaten mit gleichen Eigenschaften zu generieren.Während sich der erste Ansatz zur Erstellung neuer Datensätze eignet, kann der zweite Ansatz zur Anreicherung existierender Datensätze genutzt werden.
Schließlich liefert diese Arbeit noch zwei Beiträge zur Detektion von Angriffsszenarien. Im ersten Beitrag wird ein Konzept zur Detektion von Angriffsszenarien entwickelt, welches sich an die typischen Phasen eines Angriffsszenarios orientiert. Im zweiten Beitrag werden eine überwachte und eine unüberwachte Methode zur Detektion von langsamen Port Scans vorgestellt.
The DAEDALUS mission concept aims at exploring and characterising the entrance and initial part of Lunar lava tubes within a compact, tightly integrated spherical robotic device, with a complementary payload set and autonomous capabilities.
The mission concept addresses specifically the identification and characterisation of potential resources for future ESA exploration, the local environment of the subsurface and its geologic and compositional structure.
A sphere is ideally suited to protect sensors and scientific equipment in rough, uneven environments.
It will house laser scanners, cameras and ancillary payloads.
The sphere will be lowered into the skylight and will explore the entrance shaft, associated caverns and conduits. Lidar (light detection and ranging) systems produce 3D models with high spatial accuracy independent of lighting conditions and visible features.
Hence this will be the primary exploration toolset within the sphere.
The additional payload that can be accommodated in the robotic sphere consists of camera systems with panoramic lenses and scanners such as multi-wavelength or single-photon scanners.
A moving mass will trigger movements.
The tether for lowering the sphere will be used for data communication and powering the equipment during the descending phase.
Furthermore, the connector tether-sphere will host a WIFI access point, such that data of the conduit can be transferred to the surface relay station. During the exploration phase, the robot will be disconnected from the cable, and will use wireless communication.
Emergency autonomy software will ensure that in case of loss of communication, the robot will continue the nominal mission.
Modern immersive multimodal technologies enable the learners to completely get immersed in various learning situations in a way that feels like experiencing an authentic learning environment. These environments also allow the collection of multimodal data, which can be used with artificial intelligence to further improve the immersion and learning outcomes. The use of artificial intelligence has been widely explored for the interpretation of multimodal data collected from multiple sensors, thus giving insights to support learners’ performance by providing personalised feedback. In this paper, we present a conceptual approach for creating immersive learning environments, integrated with multi-sensor setup to help learners improve their psychomotor skills in a remote setting.
The safety of future spaceflight depends on space surveillance and space traffic management, as the density of objects in Earth orbit has reached a level that requires collision avoidance maneuvers to be performed on a regular basis to avoid a mission or, in the context of human space flight, life-endangering threat. Driven by enhanced sensor systems capable of detecting centimeter-sized debris, megaconstellations and satellite miniaturization, the space debris problem has revealed many parallels to the plastic waste in our oceans, however with much less visibility to the eye. Future catalog sizes are expected to increase drastically, making it even more important to detect potentially dangerous encounters as early as possible.
Due to the limited number of monitoring sensors, continuous observation of all objects is impossible, resulting in the need to predict the orbital paths and their uncertainty via models to perform collision risk assessment and space object catalog maintenance. For many years the uncertainty models used for orbit determination neglected any uncertainty in the astrodynamic force models, thereby implicitly assuming them to be flawless descriptions of the true space environment. This assumption is known to result in overly optimistic uncertainty estimates, which in turn complicate collision risk analysis.
The keynote of this doctoral thesis is to establish uncertainty realism for low Earth orbiting satellites via a physically connected quantification of the dominant force model uncertainties, particularly multiple sources of atmospheric density uncertainty and orbital gravity uncertainty.
The resulting process noise models are subsequently integrated into classical and state of the art orbit determination algorithms. Their positive impact is demonstrated via numerical orbit determination simulations and a collision risk assessment study using all non-restricted objects in the official United States space catalogs. It is shown that the consideration of atmospheric density uncertainty and gravity uncertainty significantly improves the quality of the orbit determination and thus makes a contribution to future spaceflight safety by increasing the reliability of the uncertainty estimates used for collision risk assessment.
To deliver the best user experience (UX), the human-centered design cycle (HCDC) serves as a well-established guideline to application developers. However, it does not yet cover network-specific requirements, which become increasingly crucial, as most applications deliver experience over the Internet. The missing network-centric view is provided by Quality of Experience (QoE), which could team up with UX towards an improved overall experience. By considering QoE aspects during the development process, it can be achieved that applications become network-aware by design. In this paper, the Quality of Experience Centered Design Cycle (QoE-CDC) is proposed, which provides guidelines on how to design applications with respect to network-specific requirements and QoE. Its practical value is showcased for popular application types and validated by outlining the design of a new smartphone application. We show that combining HCDC and QoE-CDC will result in an application design, which reaches a high UX and avoids QoE degradation.
Due to biased assumptions on the underlying ordinal rating scale in subjective Quality of Experience (QoE) studies, Mean Opinion Score (MOS)-based evaluations provide results, which are hard to interpret and can be misleading. This paper proposes to consider the full QoE distribution for evaluating, reporting, and modeling QoE results instead of relying on MOS-based metrics derived from results based on ordinal rating scales. The QoE distribution can be represented in a concise way by using the parameters of a multinomial distribution without losing any information about the underlying QoE ratings, and even keeps backward compatibility with previous, biased MOS-based results. Considering QoE results as a realization of a multinomial distribution allows to rely on a well-established theoretical background, which enables meaningful evaluations also for ordinal rating scales. Moreover, QoE models based on QoE distributions keep detailed information from the results of a QoE study of a technical system, and thus, give an unprecedented richness of insights into the end users’ experience with the technical system. In this work, existing and novel statistical methods for QoE distributions are summarized and exemplary evaluations are outlined. Furthermore, using the novel concept of quality steps, simulative and analytical QoE models based on QoE distributions are presented and showcased. The goal is to demonstrate the fundamental advantages of considering QoE distributions over MOS-based evaluations if the underlying rating data is ordinal in nature.
In many real world settings, imbalanced data impedes model performance of learning algorithms, like neural networks, mostly for rare cases. This is especially problematic for tasks focusing on these rare occurrences. For example, when estimating precipitation, extreme rainfall events are scarce but important considering their potential consequences. While there are numerous well studied solutions for classification settings, most of them cannot be applied to regression easily. Of the few solutions for regression tasks, barely any have explored cost-sensitive learning which is known to have advantages compared to sampling-based methods in classification tasks. In this work, we propose a sample weighting approach for imbalanced regression datasets called DenseWeight and a cost-sensitive learning approach for neural network regression with imbalanced data called DenseLoss based on our weighting scheme. DenseWeight weights data points according to their target value rarities through kernel density estimation (KDE). DenseLoss adjusts each data point’s influence on the loss according to DenseWeight, giving rare data points more influence on model training compared to common data points. We show on multiple differently distributed datasets that DenseLoss significantly improves model performance for rare data points through its density-based weighting scheme. Additionally, we compare DenseLoss to the state-of-the-art method SMOGN, finding that our method mostly yields better performance. Our approach provides more control over model training as it enables us to actively decide on the trade-off between focusing on common or rare cases through a single hyperparameter, allowing the training of better models for rare data points.