TY - THES A1 - Milbrandt, Jens T1 - Performance Evaluation of Efficient Resource Management Concepts for Next Generation IP Networks T1 - Effiziente Konzepte und Leistungsbewertung zum Ressourcen Management in zukünftigen IP Netzen N2 - Next generation networks (NGNs) must integrate the services of current circuit-switched telephone networks and packet-switched data networks. This convergence towards a unified communication infrastructure necessitates from the high capital expenditures (CAPEX) and operational expenditures (OPEX) due to the coexistence of separate networks for voice and data. In the end, NGNs must offer the same services as these legacy networks and, therefore, they must provide a low-cost packet-switched solution with real-time transport capabilities for telephony and multimedia applications. In addition, NGNs must be fault-tolerant to guarantee user satisfaction and to support business-critical processes also in case of network failures. A key technology for the operation of NGNs is the Internet Protocol (IP) which evolved to a common and well accepted standard for networking in the Internet during the last 25 years. There are two basically different approaches to achieve QoS in IP networks. With capacity overprovisioning (CO), an IP network is equipped with sufficient bandwidth such that network congestion becomes very unlikely and QoS is maintained most of the time. The second option to achieve QoS in IP networks is admission control (AC). AC represents a network-inherent intelligence that admits real-time traffic flows to a single link or an entire network only if enough resources are available such that the requirements on packet loss and delay can be met. Otherwise, the request of a new flow is blocked. This work focuses on resource management and control mechanisms for NGNs, in particular on AC and associated bandwidth allocation methods. The first contribution consists of a new link-oriented AC method called experience-based admission control (EBAC) which is a hybrid approach dealing with the problems inherent to conventional AC mechanisms like parameter-based or measurement-based AC (PBAC/MBAC). PBAC provides good QoS but suffers from poor resource utilization and, vice versa, MBAC uses resources efficiently but is susceptible to QoS violations. Hence, EBAC aims at increasing the resource efficiency while maintaining the QoS which increases the revenues of ISPs and postpones their CAPEX for infrastructure upgrades. To show the advantages of EBAC, we first review today’s AC approaches and then develop the concept of EBAC. EBAC is a simple mechanism that safely overbooks the capacity of a single link to increase its resource utilization. We evaluate the performance of EBAC by its simulation under various traffic conditions. The second contribution concerns dynamic resource allocation in transport networks which implement a specific network admission control (NAC) architecture. In general, the performance of different NAC systems may be evaluated by conventional methods such as call blocking analysis which has often been applied in the context of multi-service asynchronous transfer mode (ATM) networks. However, to yield more practical results than abstract blocking probabilities, we propose a new method to compare different AC approaches by their respective bandwidth requirements. To present our new method for comparing different AC systems, we first give an overview of network resource management (NRM) in general. Then we present the concept of adaptive bandwidth allocation (ABA) in capacity tunnels and illustrate the analytical performance evaluation framework to compare different AC systems by their capacity requirements. Different network characteristics influence the performance of ABA. Therefore, the impact of various traffic demand models and tunnel implementations, and the influence of resilience requirements is investigated. In conclusion, the resources in NGNs must be exclusively dedicated to admitted traffic to guarantee QoS. For that purpose, robust and efficient concepts for NRM are required to control the requested bandwidth with regard to the available transmission capacity. Sophisticated AC will be a key function for NRM in NGNs and, therefore, efficient resource management concepts like experience-based admission control and adaptive bandwidth allocation for admission-controlled capacity tunnels, as presented in this work are appealing for NGN solutions. N2 - In meiner Dissertation zum Thema “Performance Evaluation of Efficient Resource Management Concepts for Next Generation IP Networks” werden im Wesentlichen zwei miteinander verwobene Konzepte zur effizienten Nutzung von Übertragungsressourcen in zukünftigen IP Netzen untersucht. Das Management solcher Ressourcen ist zur Unterstützung qualitativ hochwertiger Netzdiensten (z.B. IP Telephonie, IP TV, etc.) in Zukunft unabdingbar. Gegenwärtig werden diese Dienste durch den Einsatz hoher Übertragungskapazitäten (engl. capacity overprovisioning) in den IP Breitbandnetzen ermöglicht. Um in Spitzenlastzeiten die Qualität der Dienste aufrecht zu erhalten, sind die Bandbreiten derart hoch angesetzt, dass unter normalen Umständen die Ressourcen nur sehr schwach ausgelastet sind (im Bereich zwischen 10 und 30 Prozent). Diese Überdimensionierungslösung ist einfach zu realisieren aber auch sehr kostenintensiv, ineffizient und vor allem nicht (zukunfts-)sicher, da bei ständig steigendem Bandbreitenbedarf, die Netzkapazitäten häufig angepasst werden müssen. Eine Effizienzsteigerung bei der Ressourcennutzung in heutigen Kommunikationsnetzen ist daher ein wichtiges Kriterium für die Wirtschaftlichkeit zukünftiger IP Netze. Erreicht werden kann dies mit den Mitteln der Netzzugangskontrolle (engl. admission control, kurz AC), welche bereits in verschiedenen Formen entwickelt, untersucht und teilweise auch in heutigen IP Netzen realisiert ist. Die AC stellt eine vergleichsweise komplexe Lösung zur Aufrechterhaltung der Dienstgüte in IP Netzen dar. Daher sind einfache und effiziente Mechanismen/Automatismen zur Durchführung der AC gefordert, um deren Einsatz an Stelle der Überdimensionierung zu rechtfertigen. Den zuvor genannten Forderungen nach effizienter Ressourcennutzung entsprechend stellt der erste Hauptbeitrag der Dissertation einen neuen Ansatz zur AC dar, die so genannte erfahrungsbasierte Netzzugangskontrolle (engl. experience-based admission control, kurz EBAC). Gegenüber den existierenden alternativen, d.h. parameter- oder mess-basierten Verfahren der AC zeichnet sich die EBAC durch effiziente Ressourcennutzung und begünstigt gleichzeitig die Aufrechterhaltung der Dienstgüte. Die genaue Funktionsweise der EBAC und die simulative Leistungsuntersuchung sind in der Dissertation nachzulesen. Der zweite Themenschwerpunkt der Dissertation greift eine tunnel-basierte Netzarchitektur auf und präsentiert mit deren Hilfe eine neue Methode zur Bewertung verschiedener Bandbreitenallokationsstrategien in Kombination mit AC. Die neue Bewertungsmethode wird zwar anhand einer speziellen tunnel-basierten AC untersucht, lässt sich aber auf alle Arten so genannter budget-basierter AC anwenden. Bei der konventionellen Leistungsbewertung verschiedener AC-Systeme dienen abstrakte Blockierungswahrscheinlichkeiten als Leistungsmaß. Im Gegensatz hierzu, bewertet die neue Methode die Leistung eines AC-Systems anhand des zugehörigen Bandbreitenbedarfs in Abhängigkeit verschiedener Einflussfaktoren, z.B. der avisierten Blockierwahrscheinlichkeit des AC-Systems, der Verkehrszusammensetzung, des Verkehrsvolumens, der Verkehrsdynamik und vor allem der angewandten Bandbreitenallokationsstrategie. In zukünftigen IP Netzen müssen die Ressourcen zur Datenübertragung besser verwaltet und kontrolliert werden. Zugelassener Verkehr mit hohen Dienstgüteanforderungen muss exklusiv Bandbreite zugewiesen werden, um die geforderte Dienstgüte garantieren zu können. Um dieses Ziel zu erreichen werden robuste und effiziente Konzepte zum Management von Netzressourcen benötigt. Intelligenter Netzzugangskontrolle wird eine Schlüsselfunktion in zukünftigen IP Netzen zukommen. Die in dieser Dissertation vorgestellten, effizienten Ressourcenmanagementkonzepte „erfahrungsbasierte Netzzugangskontrolle“ und „adaptive Bandbreitenallokation in zugangskontrollierten Kapazitätstunnel“ tragen zum Erreichen dieses Ziels bei. T3 - Würzburger Beiträge zur Leistungsbewertung Verteilter Systeme - 01/07 KW - Ressourcenmanagement KW - Leistungsbewertung KW - Rechnernetz KW - IP KW - Ressourcen Management KW - Leistungsbewertung KW - zukünftige Kommunikationsnetze KW - Resource Management KW - Performance Analysis KW - Next Generation Networks Y1 - 2007 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-23332 ER - TY - THES A1 - Atzmüller, Martin T1 - Knowledge-Intensive Subgroup Mining - Techniques for Automatic and Interactive Discovery T1 - Wissensintensive Subgruppenentdeckung – Automatische und Interaktive Entdeckungsmethoden N2 - Data mining has proved its significance in various domains and applications. As an important subfield of the general data mining task, subgroup mining can be used, e.g., for marketing purposes in business domains, or for quality profiling and analysis in medical domains. The goal is to efficiently discover novel, potentially useful and ultimately interesting knowledge. However, in real-world situations these requirements often cannot be fulfilled, e.g., if the applied methods do not scale for large data sets, if too many results are presented to the user, or if many of the discovered patterns are already known to the user. This thesis proposes a combination of several techniques in order to cope with the sketched problems: We discuss automatic methods, including heuristic and exhaustive approaches, and especially present the novel SD-Map algorithm for exhaustive subgroup discovery that is fast and effective. For an interactive approach we describe techniques for subgroup introspection and analysis, and we present advanced visualization methods, e.g., the zoomtable that directly shows the most important parameters of a subgroup and that can be used for optimization and exploration. We also describe various visualizations for subgroup comparison and evaluation in order to support the user during these essential steps. Furthermore, we propose to include possibly available background knowledge that is easy to formalize into the mining process. We can utilize the knowledge in many ways: To focus the search process, to restrict the search space, and ultimately to increase the efficiency of the discovery method. We especially present background knowledge to be applied for filtering the elements of the problem domain, for constructing abstractions, for aggregating values of attributes, and for the post-processing of the discovered set of patterns. Finally, the techniques are combined into a knowledge-intensive process supporting both automatic and interactive methods for subgroup mining. The practical significance of the proposed approach strongly depends on the available tools. We introduce the VIKAMINE system as a highly-integrated environment for knowledge-intensive active subgroup mining. Also, we present an evaluation consisting of two parts: With respect to objective evaluation criteria, i.e., comparing the efficiency and the effectiveness of the subgroup discovery methods, we provide an experimental evaluation using generated data. For that task we present a novel data generator that allows a simple and intuitive specification of the data characteristics. The results of the experimental evaluation indicate that the novel SD-Map method outperforms the other described algorithms using data sets similar to the intended application concerning the efficiency, and also with respect to precision and recall for the heuristic methods. Subjective evaluation criteria include the user acceptance, the benefit of the approach, and the interestingness of the results. We present five case studies utilizing the presented techniques: The approach has been successfully implemented in medical and technical applications using real-world data sets. The method was very well accepted by the users that were able to discover novel, useful, and interesting knowledge. N2 - Data Mining wird mit großem Erfolg in vielen Domänen angewandt. Subgruppenentdeckung als wichtiges Teilgebiet des Data Mining kann zum Beispiel gut im Marketing, oder zur Qualitätskontrolle und Analyse in medizinischen Domänen eingesetzt werden. Das allgemeine Ziel besteht darin, potentiell nützliches and letztendlich interessantes Wissen zu entdecken. Jedoch können diese Anforderungen im praktischen Einsatz oft nicht erfüllt werden, etwa falls die eingesetzten Methoden eine schlechte Skalierbarkeit für größere Datensätze aufweisen, falls dem Benutzer zu viele Ergebnisse präsentiert werden, oder falls der Anwender viele der gefundenen Subgruppen-Muster schon kennt. Diese Arbeit stellt eine Kombination von automatischen und interaktiven Techniken vor, um mit den genannten Problemen besser umgehen zu können: Es werden automatische heuristische und vollständige Subgruppenentdeckungs-Verfahren diskutiert, und insbesondere der neuartige SD-Map Algorithmus zur vollständigen Subgruppenentdeckung vorgestellt der sowohl schnell als auch effektiv ist. Bezüglich der interaktiven Techniken werden Methoden zur Subgruppen-Introspektion und Analyse, und fortgeschrittene Visualisierungstechniken vorgestellt, beispielsweise die Zoomtable, die die für die Subgruppenentdeckung wichtigsten Parameter direkt visualisiert und zur Optimierung und Exploration eingesetzt werden kann. Zusätzlich werden verschiedene Visualisierungen zum Vergleich und zur Evaluation von Subgruppen beschrieben um den Benutzer bei diesen essentiellen Schritten zu unterstützen. Weiterhin wird leicht zu formalisierendes Hintergrundwissen vorgestellt, das im Subgruppenentdeckungsprozess in vielfältiger Weise eingesetzt werden kann: Um den Entdeckungsprozess zu fokussieren, den Suchraum einzuschränken, und letztendlich die Effizienz der Entdeckungsmethode zu erhöhen. Insbesondere wird Hintergrundwissen eingeführt, um die Elemente der Anwendungsdomäne zu filtern, um geeignete Abstraktionen zu definieren, Werte zusammenzufassen, und die gefundenen Subgruppenmuster nachzubearbeiten. Schließlich werden diese Techniken in einen wissensintensiven Prozess integriert, der sowohl automatische als auch interaktive Methoden zur Subgruppenentdeckung einschließt. Die praktische Bedeutung des vorgestellten Ansatzes hängt stark von den verfügbaren Werkzeugen ab. Dazu wird das VIKAMINE System als hochintegrierte Umgebung für die wissensintensive aktive Subgruppenentdeckung präsentiert. Die Evaluation des Ansatzes besteht aus zwei Teilen: Hinsichtlich einer Evaluation von Effizienz und Effektivität der Verfahren wird eine experimentelle Evaluation mit synthetischen Daten vorgestellt. Für diesen Zweck wird ein neuartiger in der Arbeit entwickelter Datengenerator angewandt, der eine einfache und intuitive Spezifikation der Datencharakteristiken erlaubt. Für die Evaluation des Ansatzes wurden Daten erzeugt, die ähnliche Charakteristiken aufweisen wie die Daten des angestrebten Einsatzbereichs. Die Ergebnisse der Evaluation zeigen, dass der neuartige SD-Map Algorithmus den anderen in der Arbeit beschriebenen Standard-Algorithmen überlegen ist. Sowohl hinsichtlich der Effizienz, als auch von Precision/Recall bezogen auf die heuristischen Algorithmen bietet SD-Map deutliche Vorteile. Subjektive Evaluationskriterien sind durch die Benutzerakzeptanz, den Nutzen des Ansatzes, und die Interessantheit der Ergebnisse gegeben. Es werden fünf Fallstudien für den Einsatz der vorgestellten Techniken beschrieben: Der Ansatz wurde in medizinischen und technischen Anwendungen mit realen Daten eingesetzt. Dabei wurde er von den Benutzern sehr gut angenommen, und im praktischen Einsatz konnte neuartiges, nützliches, und interessantes Wissen entdeckt werden. KW - Data Mining KW - Algorithmus KW - Visualisierung KW - Subgruppenentdeckung KW - Hintergrundwissen KW - Wissensendeckung KW - Data Mining KW - Visualisierung KW - Subgroup Mining KW - Background Knowledge KW - Knowledge Discovery KW - Data Mining KW - Visualization Y1 - 2006 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-21004 ER - TY - THES A1 - Travers, Stephen T1 - Structural Properties of NP-Hard Sets and Uniform Characterisations of Complexity Classes T1 - Strukturelle Eigenschaften NP-harter Mengen und uniforme Charakterisierungen von Komplexitätsklassen N2 - This thesis is devoted to the study of computational complexity theory, a branch of theoretical computer science. Computational complexity theory investigates the inherent difficulty in designing efficient algorithms for computational problems. By doing so, it analyses the scalability of computational problems and algorithms and places practical limits on what computers can actually accomplish. Computational problems are categorised into complexity classes. Among the most important complexity classes are the class NP and the subclass of NP-complete problems, which comprises many important optimisation problems in the field of operations research. Moreover, with the P-NP-problem, the class NP represents the most important unsolved question in computer science. The first part of this thesis is devoted to the study of NP-complete-, and more generally, NP-hard problems. It aims at improving our understanding of this important complexity class by systematically studying how altering NP-hard sets affects their NP-hardness. This research is related to longstanding open questions concerning the complexity of unions of disjoint NP-complete sets, and the existence of sparse NP-hard sets. The second part of the thesis is also dedicated to complexity classes but takes a different perspective: In a sense, after investigating the interior of complexity classes in the first part, the focus shifts to the description of complexity classes and thereby to the exterior in the second part. It deals with the description of complexity classes through leaf languages, a uniform framework which allows us to characterise a great variety of important complexity classes. The known concepts are complemented by a new leaf-language model. To a certain extent, this new approach combines the advantages of the known models. The presented results give evidence that the connection between the theory of formal languages and computational complexity theory might be closer than formerly known. N2 - Diese Dissertation behandelt die Komplexitätstheorie, ein zentrales Teilgebiet der Theoretischen Informatik. Die Komplexitätstheorie untersucht die inhärente Schwierigkeit, effiziente Algorithmen für Berechnungsprobleme zu entwerfen. Sie analysiert die Skalierbarkeit von Berechnungsproblemen und Algorithmen und stellt grundsätzliche Grenzen für die Leistungsfähigkeit von Computern auf. Berechnungsprobleme werden in Komplexitätsklassen kategorisiert. Dabei spielen die Klasse NP und die in ihr enthaltene Klasse der NP-vollständigen Probleme eine wichtige Rolle. Letztere umfasst zahlreiche in der Praxis bedeutsame Probleme aus dem Bereich Operations Research. Darüber hinaus repräsentiert die Klasse NP mit dem P-NP Problem gleichfalls das wichtigste ungelöste Problem in der Informatik. Der erste Teil dieser Dissertation ist der Untersuchung NP-vollständiger und noch allgemeiner, NP-harter Mengen gewidmet. Durch eine systematische Untersuchung der Frage, wie sich partielle Modifikationen von Mengen auf deren NP-Härte auswirken, soll das Verständnis dieser wichtigen Komplexitätsklasse verbessert werden. Die Untersuchungen in diesem Bereich stehen in enger Verbindung zu wichtigen ungelösten Fragen, wie beispielsweise der Frage nach der Komplexität von Vereinigungen disjunkter NP-vollständiger Mengen und darüber hinaus der Frage nach der Existenz dünner, NP-harter Mengen. Der zweite Teil der Dissertation beschäftigt sich ebenfalls mit der Komplexitätstheorie, nimmt dabei aber eine andere Perspektive ein: Während im ersten Teil mit der Untersuchung struktureller Eigenschaften innere Aspekte von Komplexitätsklassen im Vordergrund stehen dreht es sich im zweiten Teil um die Beschreibung von Komplexitätsklassen. Dabei werden so genannte Blattsprachen verwendet, welche einen uniformen Beschreibungsmechanismus für Komplexitätsklassen darstellen. Die bestehenden Blattsprachen-Konzepte werden durch einen neuen Ansatz ergänzt, der in einem gewissen Sinne die Vorteile der bekannten Ansätze vereint. Die erzielten Ergebnisse sind Evidenz dafür, dass die Verbindung zwischen der Theorie der formalen Sprachen und der Komplexitätstheorie noch enger ist als bislang vermutet. KW - Berechnungskomplexität KW - Komplexität KW - Theoretische Informatik KW - NP-Vollständigkeit KW - Strukturelle Komplexität KW - NP-complete sets KW - structural complexity Y1 - 2007 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-27124 ER - TY - THES A1 - Feineis, Markus T1 - Wortgenaue Annotation digitalisierter mittelalterlicher Handschriften T1 - One-to-one Annotation of Digitised Medieval Manuscripts N2 - No abstract available KW - Annotation KW - Handschrift KW - Digitalisierung Y1 - 2008 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-30448 ER - TY - THES A1 - Mäder, Andreas T1 - Performance Models for UMTS 3.5G Mobile Wireless Systems T1 - Leistungsmodelle für UMTS 3.5G Mobilfunksysteme N2 - Mobile telecommunication systems of the 3.5th generation (3.5G) constitute a first step towards the requirements of an all-IP world. As the denotation suggests, 3.5G systems are not completely new designed from scratch. Instead, they are evolved from existing 3G systems like UMTS or cdma2000. 3.5G systems are primarily designed and optimized for packet-switched best-effort traffic, but they are also intended to increase system capacity by exploiting available radio resources more efficiently. Systems based on cdma2000 are enhanced with 1xEV-DO (EV-DO: evolution, data-optimized). In the UMTS domain, the 3G partnership project (3GPP) specified the High Speed Packet Access (HSPA) family, consisting of High Speed Downlink Packet Access (HSDPA) and its counterpart High Speed Uplink Packet Access (HSUPA) or Enhanced Uplink. The focus of this monograph is on HSPA systems, although the operation principles of other 3.5G systems are similar. One of the main contributions of our work are performance models which allow a holistic view on the system. The models consider user traffic on flow-level, such that only on significant changes of the system state a recalculation of parameters like bandwidth is necessary. The impact of lower layers is captured by stochastic models. This approach combines accurate modeling and the ability to cope with computational complexity. Adopting this approach to HSDPA, we develop a new physical layer abstraction model that takes radio resources, scheduling discipline, radio propagation and mobile device capabilities into account. Together with models for the calculation of network-wide interference and transmit powers, a discrete-event simulation and an analytical model based on a queuing-theoretical approach are proposed. For the Enhanced Uplink, we develop analytical models considering independent and correlated other-cell interference. N2 - Die vorliegende Arbeit beschäftigt sich mit Mobilfunksystemen der Generation 3.5 im Allgemeinen, und mit den UMTS-spezifischen Ausprägungen HSDPA (High Speed Downlink Packet Access) und HSUPA (High Speed Uplink Packet Access) bzw. Enhanced Uplink im speziellen. Es werden integrierte Systeme betrachtet, d.h. 3.5G Datenkanäle koexistieren mit "klassischen" UMTS Datenkanälen wie in den Spezifikationen von UMTS Release ´99 beschrieben. T3 - Würzburger Beiträge zur Leistungsbewertung Verteilter Systeme - 02/08 KW - Mobilfunk KW - Leistungsbewertung KW - UMTS KW - HSPA KW - Funkressourcenverwaltung KW - Modellierungstechniken KW - Netzwerkplanung KW - UMTS KW - HSPA KW - radio resource management KW - modeling techniques KW - network planning Y1 - 2008 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-32525 ER - TY - THES A1 - Tischler, German T1 - Theory and Applications of Parametric Weighted Finite Automata T1 - Theorie und Anwendungen parametrisch gewichteter endlicher Automaten N2 - Parametric weighted finite automata (PWFA) are a multi-dimensional generalization of weighted finite automata. The expressiveness of PWFA contains the expressiveness of weighted finite automata as well as the expressiveness of affine iterated function system. The thesis discusses theory and applications of PWFA. The properties of PWFA definable sets are studied and it is shown that some fractal generator systems can be simulated using PWFA and that various real and complex functions can be represented by PWFA. Furthermore, the decoding of PWFA and the interpretation of PWFA definable sets is discussed. N2 - Parametrisch gewichtete endliche Automaten (PWFA) sind eine multidimensionale Verallgemeinerung gewichteter endlicher Automaten, die sowohl die funktionale Mächtigkeit gewichteter endlicher Automaten als auch die affiner iterierter Funktionensystem umschließt. Die Arbeit diskutiert die Theorie und Anwendungen von PWFA. Es werden Eigenschaften der von PWFA definierbaren Mengen untersucht, gezeigt dass verschiedene fraktale Generatorsysteme mit PWFA simuliert werden können und dass viele praxisrelevante reelle und komplexe Funktionen und Relationen mit PWFA darstellbar sind. Ferner wird die Dekodierung von PWFA und die Interpretation PWFA-darstellbarer Mengen diskutiert. KW - Automat KW - Automata Theory Y1 - 2008 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-28145 ER - TY - THES A1 - Martin, Rüdiger T1 - Resilience, Provisioning, and Control for the Network of the Future T1 - Ausfallsicherheit, Dimensionierungsansätze und Kontrollmechanismen für das Netz der Zukunft N2 - The Internet sees an ongoing transformation process from a single best-effort service network into a multi-service network. In addition to traditional applications like e-mail,WWW-traffic, or file transfer, future generation networks (FGNs) will carry services with real-time constraints and stringent availability and reliability requirements like Voice over IP (VoIP), video conferencing, virtual private networks (VPNs) for finance, other real-time business applications, tele-medicine, or tele-robotics. Hence, quality of service (QoS) guarantees and resilience to failures are crucial characteristics of an FGN architecture. At the same time, network operations must be efficient. This necessitates sophisticated mechanisms for the provisioning and the control of future communication infrastructures. In this work we investigate such echanisms for resilient FGNs. There are many aspects of the provisioning and control of resilient FGNs such as traffic matrix estimation, traffic characterization, traffic forecasting, mechanisms for QoS enforcement also during failure cases, resilient routing, or calability concerns for future routing and addressing mechanisms. In this work we focus on three important aspects for which performance analysis can deliver substantial insights: load balancing for multipath Internet routing, fast resilience concepts, and advanced dimensioning techniques for resilient networks. Routing in modern communication networks is often based on multipath structures, e.g., equal-cost multipath routing (ECMP) in IP networks, to facilitate traffic engineering and resiliency. When multipath routing is applied, load balancing algorithms distribute the traffic over available paths towards the destination according to pre-configured distribution values. State-of-the-art load balancing algorithms operate either on the packet or the flow level. Packet level mechanisms achieve highly accurate traffic distributions, but are known to have negative effects on the performance of transport protocols and should not be applied. Flow level mechanisms avoid performance degradations, but at the expense of reduced accuracy. These inaccuracies may have unpredictable effects on link capacity requirements and complicate resource management. Thus, it is important to exactly understand the accuracy and dynamics of load balancing algorithms in order to be able to exercise better network control. Knowing about their weaknesses, it is also important to look for alternatives and to assess their applicability in different networking scenarios. This is the first aspect of this work. Component failures are inevitable during the operation of communication networks and lead to routing disruptions if no special precautions are taken. In case of a failure, the robust shortest-path routing of the Internet reconverges after some time to a state where all nodes are again reachable – provided physical connectivity still exists. But stringent availability and reliability criteria of new services make a fast reaction to failures obligatory for resilient FGNs. This led to the development of fast reroute (FRR) concepts for MPLS and IP routing. The operations of MPLS-FRR have already been standardized. Still, the standards leave some degrees of freedom for the resilient path layout and it is important to understand the tradeoffs between different options for the path layout to efficiently provision resilient FGNs. In contrast, the standardization for IP-FRR is an ongoing process. The applicability and possible combinations of different concepts still are open issues. IP-FRR also facilitates a comprehensive resilience framework for IP routing covering all steps of the failure recovery cycle. These points constitute another aspect of this work. Finally, communication networks are usually over-provisioned, i.e., they have much more capacity installed than actually required during normal operation. This is a precaution for various challenges such as network element failures. An alternative to this capacity overprovisioning (CO) approach is admission control (AC). AC blocks new flows in case of imminent overload due to unanticipated events to protect the QoS for already admitted flows. On the one hand, CO is generally viewed as a simple mechanism, AC as a more complex mechanism that complicates the network control plane and raises interoperability issues. On the other hand, AC appears more cost-efficient than CO. To obtain advanced provisioning methods for resilient FGNs, it is important to find suitable models for irregular events, such as failures and different sources of overload, and to incorporate them into capacity dimensioning methods. This allows for a fair comparison between CO and AC in various situations and yields a better understanding of the strengths and weaknesses of both concepts. Such an advanced capacity dimensioning method for resilient FGNs represents the third aspect of this work. N2 - Das Internet befindet sich gegenwärtig in einem Transformationsprozess von einem Netz mit einer einzigen best-effort Dienstklasse hin zu einem Mehr-Dienste-Netz. Zusätzlich zu herkömmlichen Anwendungen wie E-Mail, WWW oder Datenübertragung werden zukünftige Netze Dienste mit Echtzeitbedürfnissen und strikten Anforderungen an Verfügbarkeit und Zuverlässigkeit wie Voice over IP (VoIP), Videokonferenzdienste, Virtual Private Networks (VPNs) für Finanzanwendungen und andere Geschäftsanwendungen mit Echtzeitanforderungen, Tele-Medizin oder Telerobotik tragen. Daher sind die Gewährleistung von Dienstgüte und Ausfallsicherheit wesentliche Merkmale zukünftiger Netzarchitekturen. Gleichzeitig muss der Netzbetrieb effizient sein. Dies zieht den Bedarf an ausgefeilten Mechanismen für die Dimensionierung und Kontrolle ausfallsicherer Kommunikationsstrukturen nach sich. In dieser Arbeit werden solche Mechanismen, nämlich Lastaufteilung, Konzepte zur schnellen Reaktion im Fehlerfall und intelligente Ansätze zur Netzdimensionierung, untersucht. T3 - Würzburger Beiträge zur Leistungsbewertung Verteilter Systeme - 03/08 KW - Backbone-Netz KW - load balancing KW - resilience KW - fast reroute KW - admission control KW - overprovisioning Y1 - 2008 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-28497 ER - TY - THES A1 - Freiberg, Martina T1 - UI-, User-, & Usability-Oriented Engineering of Participative Knowledge-Based Systems T1 - UI-, Benutzer-, & Usability-orientierte Entwicklung partizipativer Wissensbasierter Systeme N2 - Knowledge-based systems (KBS) face an ever-increasing interest in various disciplines and contexts. Yet, the former aim to construct the ’perfect intelligent software’ continuously shifts to user-centered, participative solutions. Such systems enable users to contribute their personal knowledge to the problem solving process for increased efficiency and an ameliorated user experience. More precisely, we define non-functional key requirements of participative KBS as: Transparency (encompassing KBS status mediation), configurability (user adaptability, degree of user control/exploration), quality of the KB and UI, and evolvability (enabling the KBS to grow mature with their users). Many of those requirements depend on the respective target users, thus calling for a more user-centered development. Often, also highly expertise domains are targeted — inducing highly complex KBs — which requires a more careful and considerate UI/interaction design. Still, current KBS engineering (KBSE) approaches mostly focus on knowledge acquisition (KA) This often leads to non-optimal, little reusable, and non/little evaluated KBS front-end solutions. In this thesis we propose a more encompassing KBSE approach. Due to the strong mutual influences between KB and UI, we suggest a novel form of intertwined UI and KB development. We base the approach on three core components for encompassing KBSE: (1) Extensible prototyping, a tailored form of evolutionary prototyping; this builds on mature UI prototypes and offers two extension steps for the anytime creation of core KBS prototypes (KB + core UI) and fully productive KBS (core KBS prototype + common framing functionality). (2) KBS UI patterns, that define reusable solutions for the core KBS UI/interaction; we provide a basic collection of such patterns in this work. (3) Suitable usability instruments for the assessment of the KBS artifacts. Therewith, we do not strive for ’yet another’ self-contained KBS engineering methodology. Rather, we motivate to extend existing approaches by the proposed key components. We demonstrate this based on an agile KBSE model. For practical support, we introduce the tailored KBSE tool ProKEt. ProKEt offers a basic selection of KBS core UI patterns and corresponding configuration options out of the box; their further adaption/extension is possible on various levels of expertise. For practical usability support, ProKEt offers facilities for quantitative and qualitative data collection. ProKEt explicitly fosters the suggested, intertwined development of UI and KB. For seamlessly integrating KA activities, it provides extension points for two selected external KA tools: For KnowOF, a standard office based KA environment. And for KnowWE, a semantic wiki for collaborative KA. Therewith, ProKEt offers powerful support for encompassing, user-centered KBSE. Finally, based on the approach and the tool, we also developed a novel KBS type: Clarification KBS as a mashup of consultation and justification KBS modules. Those denote a specifically suitable realization for participative KBS in highly expertise contexts and consequently require a specific design. In this thesis, apart from more common UI solutions, we particularly also introduce KBS UI patterns especially tailored towards Clarification KBS. N2 - Das Interesse an wissensbasierten Systemen (WBS) in verschiedensten Fachdisziplinen und Anwendungskontexten wächst nach wie vor stetig. Das frühere Ziel — intelligente Software als Expertenersatz — verschiebt sich dabei allerdings kontinuierlich in Richtung partizipativer, nutzerzentrierter Anwendungen. Solche Systeme erlauben dem Benutzer, das vorhandene persönliche Hintergrundwissen in den Problemlösungsprozess mit einzubringen um die Effizienz des Systems zu steigern und die User Experience zu verbessern. Konkret definieren wir für partizipative WBS die folgenden, nichtfunktionalen Anforderungen: Transparenz (umfassende Vermittlung des Systemstatus), Konfigurierbarkeit (Anpassbarkeit an verschiedene Benutzer und Grad der Benutzerkontrolle und Explorationsmöglichkeit), Qualität sowohl der Wissensbasis als auch der Benutzeroberfläche und Fortentwickelbarkeit (Fähigkeit des WBS analog zu den Kenntnissen seiner Nutzer zu reifen). Viele dieser Anforderungen hängen stark von den jeweiligen Nutzern ab. Im Umkehrschluss erfordert dies eine nutzerzentriertere Entwicklung solcher Systeme. Die häufig sehr fachspezifischen Zieldomänen haben meist entsprechend komplexe Wissensbasen zur Folge. Dies verlangt erst Recht nach einem wohlüberlegten, durchdachten UI- und Inkteraktionsdesign. Dem zum Trotz fokussieren aktuelle WBS Entwicklungsansätze jedoch nach wie vor auf der Wissensformalisierung. Mit der Folge, dass oft keine optimalen, schlecht wiederverwendbare und nur teilweise (oder gar nicht) evaluierte WBS UI/Interaktionslösungen entstehen. Die vorliegende Dissertation schlägt einen allumfassenderen WBS Entwicklungsansatz vor. Unter Berücksichtigung der starken, wechselseitigen Beeinflussung von Wissensbasis und UI ist dieser geprägt durch eine starke Verzahnung von Wissensbasis- und UI-Entwicklung. Der Ansatz stützt sich auf drei Kernkomponenten für allumfassende Entwicklung wissensbasierter Systeme: (1) Extensible Prototyping, eine adaptierte Form des evolutionären Prototyping. Extensible Prototyping basiert auf hochentwickelten UI Prototypen und definiert zwei Erweiterungsschritte um jederzeit WBS-Kernprototypen (Wissensbasis und Kern-UI) beziehungsweise voll funktionale wissensbasierte Anwendungen (WBS Kernprototyp und Rahmenfunktionalität) zu erstellen. (2) WBS UI Patterns. Diese Patterns, oder Muster, definieren wiederverwendbare Lösungen für die Kern-UI und -Inkteraktion. In dieser Arbeit stellen wir eine Sammlung grundlegender WBS UI Patterns vor. (3) Passende Usability Techniken die sich speziell für die Evaluation von WBS Software eignen. Insgesamt streben wir keine weitere, in sich geschlossene WBS Entwicklungsmethodologie an. Vielmehr motivieren wir, existierende, Wissensformalisierungs-lastige Ansätze um die vorgeschlagenen Kernkomponenten zu erweitern. Wir demonstrieren das am agilen Prozessmodell für wissensbasierte Systeme. Für die praktische Umsetzung des vorgestellten Ansatzes stellen wir außerdem das spezialisierte WBS Prototyping- und Softwareentwicklungswerkzeug ProKEt vor. ProKEt unterstützt eine Auswahl der interessantesten WBS UI Patterns sowie zugehöriger Konfigurationsoptionen. Deren weitere Anpassung beziehungsweise Erweiterung ist auf verschiedenen Expertise-Leveln möglich. Um ebenso die Anwendung von Usability Techniken zu unterstützen, bietet ProKEt weiterhin Funktionalität für die quantitative und qualitative Datensammlung. Auch baut ProKEt bewusst auf die strikt verzahnte Entwicklung von UI- und Wissensbasis. Für die einfache und nahtlose Integration von Wissensformalisierung in den Gesamtprozess unterstützt ProKEt zwei externe Wissensformalisierungs-Werkzeuge: KnowOF, eine Wissensformalisierungsumgebung welche standartisierte Office Dokumente nutzt. Und KnowWE, ein semantisches Wiki für die kollaborative Wissensformalisierung im Web. Damit ist ProKEt ein mächtiges Werkzeug für umfassende, nutzerzentrierte WBS Entwicklung. Mithilfe des Entwicklungsansatzes und des Werkzeugs ProKEt haben wir weiterhin einen neuartigen WBS Typ entwickelt: Wissensbasierte Klärungssysteme, im Wesentlichen ein Mashup aus Beratungs-, Erklärungs- und Rechtfertigungskomponente. Diese Systeme stellen eine angepasste Realisierung von partizipativen WBS für höchst fachliche Kontexte dar und verlangten entsprechend nach einem sehr speziellen Design. In der vorliegenden Arbeit stellen wir daher neben allgemeinen WBS UI Patterns auch einige spezialisierte Varianten für wissensbasierte Klärungssysteme vor. KW - Wissensbasiertes System KW - Knowledge-based Systems Engineering KW - Expert System KW - Usability KW - UI and Interaction Design KW - User Participation KW - Expertensystem Y1 - 2015 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-106072 SN - 978-3-95826-012-2 (print) SN - 978-3-95826-013-9 (online) PB - Würzburg University Press ER - TY - THES A1 - Kaußner, Armin T1 - Dynamische Szenerien in der Fahrsimulation T1 - Dynamic scenarios for driving simulation N2 - In der Arbeit wird ein neues Konzept für Fahrsimulator-Datenbasen vorgestellt. Der Anwender entwirft eine auf seine Fragestellung zugeschnittene Datenbasis mithilfe einer einfachen Skriptsprache. Das Straßennetzwerk wird auf einer topologischen Ebene repäsentiert. In jedem Simulationsschritt wird hieraus im Sichtbarkeitsbereich des Fahrers die geometrische Repäsentation berechnet. Die für den Fahrer unsichtbaren Teile des Straßenetzwerks können während der Simulation verändert werden. Diese Veränderungen können von der Route des Fahrers oder von den in der Simulation erhobenen Messerten abhängen. Zudem kann der Anwender das Straßennetzwerk interaktiv verändern. Das vorgestellte Konzept bietet zahlreiche Möglichkeiten zur Erzeugung reproduzierbarer Szenarien für Experimente in Fahrsimulatoren. N2 - This work presents a new concept for driving simulator databases. Using a simple scripting language the user defines a database tailored for his experiment. The road network is represented in a topological way. Through this the geometrical representation is computed during the simulation in a small area surrounding the driver, including all that is visible for the driver. The parts of the road network that are not visible for the driver can be changed during simulation. This modification can depend on the route the driver takes or on measures available in the simulation. Moreover, the user can change the road network interactively. The presented concept offers various advantages for the design of reproducible scenarios in driving simulators. KW - Straßenverkehr KW - Simulation KW - Fahrsimulator KW - Fahrsimulation KW - Datenbasis KW - Straßennetzwerk KW - Szenariogenerierung KW - driving simulation KW - database KW - road network KW - scenario creation Y1 - 2003 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-8286 ER - TY - THES A1 - Karch, Oliver T1 - Where am I? - Indoor localization based on range measurements T1 - Wo bin ich? - Lokalisation mit Hilfe von Entfernungsmesswerten N2 - Nowadays, robotics plays an important role in increasing fields of application. There exist many environments or situations where mobile robots instead of human beings are used, since the tasks are too hazardous, uncomfortable, repetitive, or costly for humans to perform. The autonomy and the mobility of the robot are often essential for a good solution of these problems. Thus, such a robot should at least be able to answer the question "Where am I?". This thesis investigates the problem of self-localizing a robot in an indoor environment using range measurements. That is, a robot equipped with a range sensor wakes up inside a building and has to determine its position using only its sensor data and a map of its environment. We examine this problem from an idealizing point of view (reducing it into a pure geometric one) and further investigate a method of Guibas, Motwani, and Raghavan from the field of computational geometry to solving it. Here, so-called visibility skeletons, which can be seen as coarsened representations of visibility polygons, play a decisive role. In the major part of this thesis we analyze the structures and the occurring complexities in the framework of this scheme. It turns out that the main source of complication are so-called overlapping embeddings of skeletons into the map polygon, for which we derive some restrictive visibility constraints. Based on these results we are able to improve one of the occurring complexity bounds in the sense that we can formulate it with respect to the number of reflex vertices instead of the total number of map vertices. This also affects the worst-case bound on the preprocessing complexity of the method. The second part of this thesis compares the previous idealizing assumptions with the properties of real-world environments and discusses the occurring problems. In order to circumvent these problems, we use the concept of distance functions, which model the resemblance between the sensor data and the map, and appropriately adapt the above method to the needs of realistic scenarios. In particular, we introduce a distance function, namely the polar coordinate metric, which seems to be well suited to the localization problem. Finally, we present the RoLoPro software where most of the discussed algorithms are implemented (including the polar coordinate metric). N2 - Heutzutage spielen autonome Roboter bei einer wachsenden Zahl von Anwendungsgebieten eine entscheidende Rolle. Sie werden überall dort anstelle von menschlichen Arbeitskräften eingesetzt, wo die jeweiligen Aufgaben für Menschen zu gefährlich, unangenehm, monoton oder schlicht zu teuer sind. Dabei sind die Autonomie und Mobilität des Roboters sehr oft grundlegend für eine gute Problemlösung. Ein solcher Roboter sollte also zumindest die Frage "Wo bin ich?" zufriedenstellend beantworten können. Diese Arbeit behandelt das Problem der Selbstlokalisation in einer Gebäudeumgebung mit Hilfe von Entfernungsmesswerten. Das heißt, ein Roboter - ausgestattet mit einem Entfernungssensor - wacht innerhalb eines Gebäudes auf und muss mit Hilfe seiner Sensordaten und einer Karte seiner Einsatzumgebung seine Position bestimmen. Wir betrachten eine idealisierte Variante dieser Aufgabe, die ein rein geometrisches Problem zum Inhalt hat, und untersuchen ein Verfahren von Guibas, Motwani und Raghavan aus dem Gebiet der Algorithmischen Geometrie, welches dieses löst. Hierbei spielen sogenannte Sichtbarkeitsskelette (vergröberte Darstellungen von Sichtbarkeitspolygonen) eine entscheidende Rolle. Im Hauptteil der Arbeit analysieren wir die Strukturen und die auftretenden Komplexitäten im Rahmen dieses Verfahrens. Es stellt sich heraus, dass die Hauptschwierigkeiten sogenannte überlappende Einbettungen von Skeletten in das Kartenpolygon zur Ursache haben, für die wir einige einschränkende Sichtbarkeitsbedingungen zeigen. Gestützt auf diese Resultate können wir die auftretenden Komplexitätsschranken dahingehend verbessern, dass wir diese nicht nur in Abhängigkeit der Gesamtzahl aller Kartenecken angeben, sondern in Abhängigkeit der Zahl der konkaven Ecken. Dies hat ebenfalls Auswirkungen auf die Worst-Case-Schranken für die Preprocessing-Komplexität des Verfahrens. Der zweite Teil der Arbeit vergleicht die anfangs gemachten idealisierenden Annahmen mit den Gegebenheiten realer Umgebungen und adressiert die auftretenden Probleme. Um diese zu umgehen verwenden wir das Konzept sogenannter Distanzfunktionen, welche die Ähnlichkeit zwischen den Sensordaten und der Karte modellieren, und passen das Verfahren auf geeignete Weise an die Bedürfnisse realistischer Szenarien an. Insbesondere führen wir eine Distanzfunktion ein - die Polarkoordinatenmetrik - welche sich für das Lokalisationsproblem besonders gut zu eignen scheint. Schlussendlich stellen wir die Software RoLoPro vor, in der die meisten der diskutierten Algorithmen (einschließlich der Polarkoordinatenmetrik) implementiert sind. KW - Autonomer Roboter KW - Mobiler Roboter KW - Lokalisation KW - Lokalisation KW - Autonomer Roboter KW - Sichtbarkeit KW - Ähnlichkeitsmaß KW - Algorithmische Geometrie KW - Localization KW - Autonomous Robot KW - Visibility KW - Similarity Measure KW - Computational Geometry Y1 - 2002 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-8442 ER - TY - RPRT A1 - Rossa, Christian T1 - Inside 2004: Multimedia und Services T1 - Inside 2004: Multimedia and Services N2 - Einzelne Artikel zu den Themen Multimedia, Dienstleistungen, IT-Sicherheit und Hochschulnetz des Rechenzentrums der Universität Würzburg N2 - Some articles concerning multimedia, services, IT security and university network in the computer center of the University of Wuerzburg KW - Multimedia KW - Dienstleistungen KW - IT-Sicherheit KW - Hochschulnetz KW - Multimedia KW - Services KW - IT Security KW - university network Y1 - 2004 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-8616 ER - TY - THES A1 - Wolz, Frank T1 - Ein generisches Konzept zur Modellierung und Bewertung feldprogrammierbarer Architekturen T1 - A generic concept for modelling and evaluating field-programmable architectures N2 - Gegenstand der Arbeit stellt eine erstmalig unternommene, architekturübergreifende Studie über feldprogrammierbare Logikbausteine zur Implementierung synchroner Schaltkreise dar. Zunächst wird ein Modell für allgemeine feldprogrammiebare Architekturen basierend auf periodischen Graphen definiert. Schließlich werden Bewertungsmaße für Architekturen und Schaltkreislayouts angegeben zur Charakterisierung struktureller Eigenschaften hinsichtlich des Verhaltens in Chipflächenverbrauch und Signalverzögerung. Ferner wird ein generisches Layout-Werkzeug entwickelt, das für beliebige Architekturen und Schaltkreise Implementierungen berechnen und bewerten kann. Abschließend werden neun ressourcenminimalistische Architekturen mit Maschen- und mit Inselstruktur einander gegenübergestellt. N2 - This work presents a first architecture-spreading study on field-programmable logical devices leaving the beaten tracks of commercial architecture improvements. After a formal model for general field-programmable architectures based on periodic graphs has been given, some feasible evaluation metrics for architectures and circuit layouts are defined characterizing structural properties of architectures in respect of chip area usage and performance. Then, a generic layout tool is developped working on arbitrary architecures and circuits. Finally, nine resource minimal mesh- and island-style architectures are compared. KW - Gay-Array-Bauelement KW - Programmierbare logische Anordnung KW - Field programmable gate array KW - Feldprogrammierbare Architekturen KW - Field-programmable Gate Arrays KW - field-programmable architectures KW - field-programmable gate arrays Y1 - 2003 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-7944 ER - TY - THES A1 - Höhn, Holger T1 - Multimediale, datenbankgestützte Lehr- und Lernplattformen T1 - Multimedia, Database Supported Teaching and Learning Environments N2 - Die Dissertation befaßt sich mit der Entwicklung einer multimedialen, datenbankgestützten Lehr- und Lernplattform. Die entwickelten Module ermöglichen und erweitern nicht nur die Möglichkeit des Selbststudiums für den Studenten sondern erleichtern auch die Arbeit der Dozenten. Außerdem wird auch die Zusammenarbeit und der Austausch von Lernobjekten zwischen verschiedenen Institutionen ermöglicht. In der Lehr- und Lernplattform können verschiedene Lernobjekt-Typen verwaltet werden. Exemplarisch wurden die Typen Bilder, 3D-Animationen, Vorlesungen, Lerntexte, Fallbeispiele und Quizelemente integriert. Die Lehr- und Lernplattform besteht aus drei Bausteinen: 1. In der Lernobjekt-Datenbank werden alle Lernobjekt-Typen und Lernobjekte verwaltet. 2. Autorenwerkzeuge dienen zur Erstellung von Lernobjekten. 3. In der Lernplattform werden die Lernobjekte den Studenten zum (Selbst-)Lernen präsentiert. Neben den Vorteilen, die der Einsatz von E-Learning im allgemeinen bietet, wie die flexible Lernorganisation oder die Nutzung von Lerninhalten unabhängig von Ort und Zeit, zeichnet sich die entwickelte Lehr- und Lernplattform besonders durch folgende Punkte aus: Generierung von Lerninhalten höherer Qualität durch multizentrische Expertenbündelung und Arbeitsteilung, Erweiterbarkeit auf andere, neue Lernobjekt-Typen, Verwaltbarkeit, Konsistenz, Flexibilität, geringer Verwaltungsaufwand, Navigationsmöglichkeiten für den Studenten, Personalisierbarkeit und Konformität zu internationalen Standards. Sowohl bei der Modellierung als auch bei der Umsetzung wurde darauf geachtet, möglichst gut die Anforderungen der Dermatologie bei gleichzeitiger Erweiterbarkeit auf andere, ähnliche Szenarien zu erfüllen. Besonders einfach sollte die Anpassung der Plattform für andere bildorientierte Disziplinen sein. N2 - In this thesis a multimedia teaching and learning environment was developed. The center of interest are learning objects which can be classified into different types. The types and objects are stored and adminstered in a database. As examples lectures, learning texts, 3D-animations, images, videos and quiz elements were integrated. For the generation of learning objects authoring tools were developed for each type. The teaching and learning environment offers the following features: lectureres are supported in preparing lectures. Because learning objects can be exchanged between different environments, the generation of "high quality" objects is encouraged. New types of learning objects can be integrated easily. Students are able to learn time-independent. The system was tested in dermatology, but the adaption to other, especially visually oriented disciplines is easy. KW - Multimedia KW - Anwendung KW - Forschung KW - Lehre KW - Lehre KW - Multimedia KW - Datenbanken KW - Lernen KW - Multimedia KW - Learning KW - Teaching KW - Databases Y1 - 2002 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-4049 ER - TY - GEN A1 - Rossa, Christian T1 - Inside 2003: IT-Sicherheit T1 - Inside 2003: IT Security N2 - Einzelne Artikel zum Thema IT-Sicherheit an der Universität Würzburg N2 - Some articles concerning IT security at the University of Wuerzburg KW - Datenübertragung ; Datensicherung ; Informationstechnik ; Internet ; Computersicherheit KW - IT-Sicherheit KW - Firewall KW - Computervirus KW - Spam-Mail KW - Webmail-System KW - IT Security KW - firewall KW - computer virus KW - spam mail KW - webmail system Y1 - 2003 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-6061 ER - TY - THES A1 - Klein, Dominik Werner T1 - Design and Evaluation of Components for Future Internet Architectures T1 - Entwurf und Bewertung von Komponenten für zukünftige Internet Architekturen N2 - Die derzeitige Internetarchitektur wurde nicht in einem geplanten Prozess konzipiert und entwickelt, sondern hat vielmehr eine evolutionsartige Entwicklung hinter sich. Auslöser für die jeweiligen Evolutionsschritte waren dabei meist aufstrebende Anwendungen, welche neue Anforderungen an die zugrundeliegende Netzarchitektur gestellt haben. Um diese Anforderungen zu erfüllen, wurden häufig neuartige Dienste oder Protokolle spezifiziert und in die bestehende Architektur integriert. Dieser Prozess ist jedoch meist mit hohem Aufwand verbunden und daher sehr träge, was die Entwicklung und Verbreitung innovativer Dienste beeinträchtigt. Derzeitig diskutierte Konzepte wie Software-Defined Networking (SDN) oder Netzvirtualisierung (NV) werden als eine Möglichkeit angesehen, die Altlasten der bestehenden Internetarchitektur zu lösen. Beiden Konzepten gemein ist die Idee, logische Netze über dem physikalischen Substrat zu betreiben. Diese logischen Netze sind hochdynamisch und können so flexibel an die Anforderungen der jeweiligen Anwendungen angepasst werden. Insbesondere erlaubt das Konzept der Virtualisierung intelligentere Netzknoten, was innovative neue Anwendungsfälle ermöglicht. Ein häufig in diesem Zusammenhang diskutierter Anwendungsfall ist die Mobilität sowohl von Endgeräten als auch von Diensten an sich. Die Mobilität der Dienste wird hierbei ausgenutzt, um die Zugriffsverzögerung oder die belegten Ressourcen im Netz zu reduzieren, indem die Dienste zum Beispiel in für den Nutzer geographisch nahe Datenzentren migriert werden. Neben den reinen Mechanismen bezüglich Dienst- und Endgerätemobilität sind in diesem Zusammenhang auch geeignete Überwachungslösungen relevant, welche die vom Nutzer wahrgenommene Dienstgüte bewerten können. Diese Lösungen liefern wichtige Entscheidungshilfen für die Migration oder überwachen mögliche Effekte der Migration auf die erfahrene Dienstgüte beim Nutzer. Im Falle von Video Streaming ermöglicht ein solcher Anwendungsfall die flexible Anpassung der Streaming Topologie für mobile Nutzer, um so die Videoqualität unabhängig vom Zugangsnetz aufrechterhalten zu können. Im Rahmen dieser Doktorarbeit wird der beschriebene Anwendungsfall am Beispiel einer Video Streaming Anwendung näher analysiert und auftretende Herausforderungen werden diskutiert. Des Weiteren werden Lösungsansätze vorgestellt und bezüglich ihrer Effizienz ausgewertet. Im Detail beschäftigt sich die Arbeit mit der Leistungsanalyse von Mechanismen für die Dienstmobilität und entwickelt eine Architektur zur Optimierung der Dienstmobilität. Im Bereich Endgerätemobilität werden Verbesserungen entwickelt, welche die Latenz zwischen Endgerät und Dienst reduzieren oder die Konnektivität unabhängig vom Zugangsnetz gewährleisten. Im letzten Teilbereich wird eine Lösung zur Überwachung der Videoqualität im Netz entwickelt und bezüglich ihrer Genauigkeit analysiert. N2 - Today’s Internet architecture was not designed from scratch but was driven by new services that emerged during its development. Hence, it is often described as patchwork where additional patches are applied in case new services require modifications to the existing architecture. This process however is rather slow and hinders the development of innovative network services with certain architecture or network requirements. Currently discussed technologies like Software-Defined Networking (SDN) or Network Virtualization (NV) are seen as key enabling technologies to overcome this rigid best effort legacy of the Internet. Both technologies offer the possibility to create virtual networks that accommodate the specific needs of certain services. These logical networks are operated on top of a physical substrate and facilitate flexible network resource allocation as physical resources can be added and removed depending on the current network and load situation. In addition, the clear separation and isolation of networks foster the development of application-aware networks that fulfill the special requirements of emerging applications. A prominent use case that benefits from these extended capabilities of the network is denoted with service component mobility. Services hosted on Virtual Machines (VMs) follow their consuming mobile endpoints, so that access latency as well as consumed network resources are reduced. Especially for applications like video streaming, which consume a large fraction of the available resources, is this an important means to relieve the resource constraints and eventually provide better service quality. Service and endpoint mobility both allow an adaptation of the used paths between an offered service, i.e., video streaming and the consuming users in case the service quality drops due to network problems. To make evidence-based adaptations in case of quality drops, a scalable monitoring component is required that is able to monitor the service quality for video streaming applications with reliable accuracy. This monograph details challenges that arise when deploying a certain service, i.e., video streaming, in a future virtualized network architecture and discusses possible solutions. In particular, this work evaluates the performance of mechanisms enabling service mobility and presents an optimized architecture for service mobility. Concerning endpoint mobility, improvements are developed that reduce the latency between endpoints and consumed services and ensure connectivity regardless of the used mobile access network. In the last part, a network-based video quality monitoring solution is developed and its accuracy is evaluated. T3 - Würzburger Beiträge zur Leistungsbewertung Verteilter Systeme - 01/14 KW - Leistungsbewertung KW - Netzwerkmanagement KW - Virtuelles Netzwerk KW - Mobiles Internet KW - Service Mobility KW - Endpoint Mobility KW - Video Quality Monitoring Y1 - 2014 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-93134 SN - 1432-8801 ER - TY - THES A1 - Fink, Martin T1 - Crossings, Curves, and Constraints in Graph Drawing T1 - Kreuzungen, Kurven und Constraints beim Zeichnen von Graphen N2 - In many cases, problems, data, or information can be modeled as graphs. Graphs can be used as a tool for modeling in any case where connections between distinguishable objects occur. Any graph consists of a set of objects, called vertices, and a set of connections, called edges, such that any edge connects a pair of vertices. For example, a social network can be modeled by a graph by transforming the users of the network into vertices and friendship relations between users into edges. Also physical networks like computer networks or transportation networks, for example, the metro network of a city, can be seen as graphs. For making graphs and, thereby, the data that is modeled, well-understandable for users, we need a visualization. Graph drawing deals with algorithms for visualizing graphs. In this thesis, especially the use of crossings and curves is investigated for graph drawing problems under additional constraints. The constraints that occur in the problems investigated in this thesis especially restrict the positions of (a part of) the vertices; this is done either as a hard constraint or as an optimization criterion. N2 - Viele Probleme, Informationen oder Daten lassen sich mit Hilfe von Graphen modellieren. Graphen können überall dort eingesetzt werden, wo Verbindungen zwischen unterscheidbaren Objekten auftreten. Ein Graph besteht aus einer Menge von Objekten, genannt Knoten, und einer Menge von Verbindungen, genannt Kanten, zwischen je einem Paar von Knoten. Ein soziales Netzwerk lässt sich etwa als Graph modellieren, indem die teilnehmenden Personen als Knoten und Freundschaftsbeziehungen als Kanten dargestellt werden. Physikalische Netzwerke wie etwa Computernetze oder Transportnetze - wie beispielsweise das U-Bahnliniennetz einer Stadt - lassen sich ebenfalls als Graph auffassen. Um Graphen und die damit modellierten Daten gut erfassen zu können benötigen wir eine Visualisierung. Das Graphenzeichnen befasst sich mit dem Entwickeln von Algorithmen zur Visualisierung von Graphen. Diese Dissertation beschäftigt sich insbesondere mit dem Einsatz von Kreuzungen und Kurven beim Zeichnen von Graphen unter Nebenbedingungen (Constraints). Die in den untersuchten Problemen auftretenden Nebenbedingungen sorgen unter anderem dafür, dass die Lage eines Teils der Knoten - als feste Anforderung oder als Optimierungskriterium - vorgegeben ist. KW - Graphenzeichnen KW - Kreuzung KW - Kurve KW - Graph KW - graph drawing KW - crossing minimization KW - curves KW - labeling KW - metro map KW - Kreuzungsminimierung KW - Landkartenbeschriftung KW - U-Bahnlinienplan Y1 - 2014 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-98235 SN - 978-3-95826-002-3 (print) SN - 978-3-95826-003-0 (online) PB - Würzburg University Press ER - TY - THES A1 - Jarschel, Michael T1 - An Assessment of Applications and Performance Analysis of Software Defined Networking T1 - Eine Untersuchung von Anwendungen und Leistungsbewertung von Software Defined Networking N2 - With the introduction of OpenFlow by the Stanford University in 2008, a process began in the area of network research, which questions the predominant approach of fully distributed network control. OpenFlow is a communication protocol that allows the externalization of the network control plane from the network devices, such as a router, and to realize it as a logically-centralized entity in software. For this concept, the term "Software Defined Networking" (SDN) was coined during scientific discourse. For the network operators, this concept has several advantages. The two most important can be summarized under the points cost savings and flexibility. Firstly, it is possible through the uniform interface for network hardware ("Southbound API"), as implemented by OpenFlow, to combine devices and software from different manufacturers, which increases the innovation and price pressure on them. Secondly, the realization of the network control plane as a freely programmable software with open interfaces ("Northbound API") provides the opportunity to adapt it to the individual circumstances of the operator's network and to exchange information with the applications it serves. This allows the network to be more flexible and to react more quickly to changing circumstances as well as transport the traffic more effectively and tailored to the user’s "Quality of Experience" (QoE). The approach of a separate network control layer for packet-based networks is not new and has already been proposed several times in the past. Therefore, the SDN approach has raised many questions about its feasibility in terms of efficiency and applicability. These questions are caused to some extent by the fact that there is no generally accepted definition of the SDN concept to date. It is therefore a part of this thesis to derive such a definition. In addition, several of the open issues are investigated. This Investigations follow the three aspects: Performance Evaluation of Software Defined Networking, applications on the SDN control layer, and the usability of SDN Northbound-API for creation application-awareness in network operation. Performance Evaluation of Software Defined Networking: The question of the efficiency of an SDN-based system was from the beginning one of the most important. In this thesis, experimental measurements of the performance of OpenFlow-enabled switch hardware and control software were conducted for the purpose of answering this question. The results of these measurements were used as input parameters for establishing an analytical model of the reactive SDN approach. Through the model it could be determined that the performance of the software control layer, often called "Controller", is crucial for the overall performance of the system, but that the approach is generally viable. Based on this finding a software for analyzing the performance of SDN controllers was developed. This software allows the emulation of the forwarding layer of an SDN network towards the control software and can thus determine its performance in different situations and configurations. The measurements with this software showed that there are quite significant differences in the behavior of different control software implementations. Among other things it has been shown that some show different characteristics for various switches, in particular in terms of message processing speed. Under certain circumstances this can lead to network failures. Applications on the SDN control layer: The core piece of software defined networking are the intelligent network applications that operate on the control layer. However, their development is still in its infancy and little is known about the technical possibilities and their limitations. Therefore, the relationship between an SDN-based and classical implementation of a network function is investigated in this thesis. This function is the monitoring of network links and the traffic they carry. A typical approach for this task has been built based on Wiretapping and specialized measurement hardware and compared with an implementation based on OpenFlow switches and a special SDN control application. The results of the comparison show that the SDN version can compete in terms of measurement accuracy for bandwidth and delay estimation with the traditional measurement set-up. However, a compromise has to be found for measurements below the millisecond range. Another question regarding the SDN control applications is whether and how well they can solve existing problems in networks. Two programs have been developed based on SDN in this thesis to solve two typical network issues. Firstly, the tool "IPOM", which enables considerably more flexibility in the study of effects of network structures for a researcher, who is confined to a fixed physical test network topology. The second software provides an interface between the Cloud Orchestration Software "OpenNebula" and an OpenFlow controller. The purpose of this software was to investigate experimentally whether a pre-notification of the network of an impending relocation of a virtual service in a data center is sufficient to ensure the continuous operation of that service. This was demonstrated on the example of a video service. Usability of the SDN Northbound API for creating application-awareness in network operation: Currently, the fact that the network and the applications that run on it are developed and operated separately leads to problems in network operation. SDN offers with the Northbound-API an open interface that enables the exchange between information of both worlds during operation. One aim of this thesis was to investigate whether this interface can be exploited so that the QoE experienced by the user can be maintained on high level. For this purpose, the QoE influence factors were determined on a challenging application by means of a subjective survey study. The application is cloud gaming, in which the calculation of video game environments takes place in the cloud and is transported via video over the network to the user. It was shown that apart from the most important factor influencing QoS, i.e., packet loss on the downlink, also the type of game type and its speed play a role. This demonstrates that in addition to QoS the application state is important and should be communicated to the network. Since an implementation of such a state conscious SDN for the example of Cloud Gaming was not possible due to its proprietary implementation, in this thesis the application “YouTube video streaming” was chosen as an alternative. For this application, status information is retrievable via the "Yomo" tool and can be used for network control. It was shown that an SDN-based implementation of an application-aware network has distinct advantages over traditional network management methods and the user quality can be obtained in spite of disturbances. N2 - Mit der Vorstellung von OpenFlow durch die Stanford Universität im Jahre 2008 begann ein Prozess im Bereich der Netzwerkforschung, der den vorherrschenden Ansatz der völlig verteilten Netzsteuerung in Frage stellt. Bei OpenFlow handelt es sich um ein Kommunikationsprotokoll, das es ermöglicht die Netzsteuerung aus den Netzwerkgeräten, z. B. einem Router, herauszulösen und als logisch-zentralisierte Einheit als reine Software zu realisieren. Ein Konzept für das im Verlauf des wissenschaftlichen Diskurses der Begriff „Software Defined Networking“ (SDN) geprägt wurde. Für die Betreiber von Netzwerken hat dieses Konzept verschiedene Vorteile. Die beiden wichtigsten lassen sich unter den Punkten Kostenersparnis und Flexibilität zusammenfassen. Zum einen ist es durch die einheitliche Schnittstelle zur Netzwerkhardware („Southbound-API“), wie sie von OpenFlow realisiert wird, möglich Geräte und Software verschiedener Hersteller miteinander zu kombinieren, was den Innovations- und Preisdruck auf diese erhöht. Zum anderen besteht durch die Realisierung der Netzwerksteuerung als frei programmierbare Software mit offenen Schnittstellen („Northbound-API“) die Möglichkeit diese an die individuellen Gegebenheiten des Betreibernetzes anzupassen als auch Informationen mit den Applikationen auszutauschen, die es bedient. Dadurch kann das Netz viel flexibler und schneller auf sich ändernde Gegebenheiten reagieren und den Verkehr effektiver und auf die vom Nutzer erfahrene „Quality of Experience“ (QoE) abgestimmt transportieren. Der Ansatz einer von der Hardware getrennten Netzwerkkontrollschicht für paket-basierte Netze ist nicht neu und wurde in der Vergangenheit bereits mehrfach vorgeschlagen. Daher sah und sieht sich der Ansatz von SDN vielen Fragen nach seiner Realisierbarkeit im Bezug auf Leistungsfähigkeit und Anwendbarkeit gegenüber. Diese Fragen rühren zum Teil auch daher, dass es bis dato keine allgemein anerkannte Definition des SDN Begriffes gibt. Daher ist es ein Teil dieser Doktorarbeit eine solche Definition herzuleiten. Darüber hinaus werden verschiedene der offenen Fragen auf ihre Lösung hin untersucht. Diese Untersuchungen folgen drei Teilaspekten: Leistungsbewertung von Software Defined Networking, Anwendungen auf der SDN Kontrollschicht und die Nutzbarkeit der SDN Northbound-API zum Herstellen eines Applikationsbewusstseins im Netzbetrieb. Leistungsbewertung von Software Defined Networking: Die Frage nach der Leistungsfähigkeit eines SDN-basierten Systems war von Beginn an eine der wichtigsten. In dieser Doktorarbeit wurden zu diesem Zweck experimentelle Messungen zur Leistung OpenFlow-fähiger Switch Hardware und Kontrollsoftware durchgeführt. Die Ergebnisse dieser Messungen dienten als Eingabeparameter zur Aufstellung eines analytischen Modells des reaktiven SDN Ansatzes. Durch das Modell ließ sich bestimmen, dass die Leistungsfähigkeit der Software Kontrollschicht, oft „Controller“ genannt, von entscheidender Bedeutung für die Gesamtleistung des Systems ist, aber der Ansatz insgesamt tragfähig ist. Basierend auf dieser Erkenntnis wurde eine Software zur Analyse der Leistungsfähigkeit von SDN Controllern entwickelt. Diese Software ermöglicht die Emulation der Weiterleitungsschicht eines SDN Netzes gegenüber der Kontrollsoftware und kann so deren Leistungsdaten in verschiedenen Situationen und Konfigurationen bestimmen. Die Messungen mit dieser Software zeigten, dass es durchaus gravierende Unterschiede im Verhalten verschiedener Kontrollsoftware Implementierungen gibt. Unter anderem konnte gezeigt werden, dass einige gegenüber verschiedenen Switches ein unterschiedliches Verhalten aufweisen, insbesondere in der Abarbeitungsgeschwindigkeit von Nachrichten. Ein Umstand der in bestimmten Fällen zu Netzausfällen führen kann. Anwendungen auf der SDN Kontrollschicht: Das Kernstück von Software Defined Networking sind die intelligenten Netzwerkapplikationen, die auf der Kontrollschicht betrieben werden. Allerdings steht deren Entwicklung noch am Anfang und es ist wenig über die technischen Möglichkeiten und deren Grenzen bekannt. Daher wurde in dieser Doktorarbeit der Frage nachgegangen, in welchem Verhältnis sich die Realisierung einer Netzwerkaufgabe mit SDN zur klassischen Umsetzung dieser bewegt. Bei dieser Aufgabe handelt es sich im speziellen um das Monitoring von Links und dem darauf befindlichen Netzwerkverkehr. Hier wurde ein typischer Ansatz für diese Aufgabe basierend auf Wiretapping und spezieller Messhardware aufgebaut und mit einer Implementierung durch OpenFlow Switches und einer speziellen SDN Kontrollapplikation verglichen. Die Ergebnisse des Vergleiches zeigen, dass die SDN Variante im Bezug auf Bandbreiten und Verzögerungsabschätzung durchaus mit dem traditionellen Messaufbau im Bezug auf Messgenauigkeit mithalten kann. Jedoch müssen Abstriche bezüglich der Verzögerungsmessung unterhalb des Millisekundenbereiches gemacht werden. Eine weitere Frage bezüglich der SDN Kontrollapplikationen besteht darin, ob und wie gut sich bestehende Probleme in Netzwerken jetzt mit SDN lösen lassen. Diesbezüglich wurden in dieser Doktorarbeit zwei Programme auf SDN Basis entwickelt, die zwei typische Netzwerkprobleme lösen. Zum einen das Tool „IPOM“, dass es ermöglicht via SDN auf Basis einer festen physikalischen Testnetz-Topologie eine andere zu emulieren und so einem Forscher auf dem Testnetz deutlich mehr Flexibilität bei der Untersuchung von Auswirkungen anderer Netzstrukturen zu ermöglichen. Die zweite Software stellt eine Schnittstelle zwischen der Cloud Orchestration Software „OpenNebula“ und einem OpenFlow Controller dar. Der Zweck dieser Software war es experimentell zu untersuchen, ob eine Vorwarnung des Netzes vor einem bevorstehenden Umzug eines virtuellen Dienstes in einem Datenzentrum hinreichend ist, um den kontinuierlichen Betrieb dieses Dienstes zu gewährleisten. Dieses konnte am Beispiel eines Videodienstes demonstriert werden. Nutzbarkeit der SDN Northbound-API zum Herstellen eines Applikationsbewusstseins im Netzbetrieb: Aktuell führt die Tatsache, dass Netz und Applikationen, die darauf laufen, separat entwickelt und betrieben werden zu Problemen im Netzbetrieb. SDN bietet mit der Northbound-API eine Schnittstelle an, die den Austausch zwischen Informationen beider Welten im Betrieb ermöglicht. Ein Ziel dieser Doktorarbeit war es zu untersuchen, ob sich diese Schnittstelle ausnutzen lässt, so dass die vom Nutzer erfahrene QoE hoch gehalten werden kann. Zu diesem Zwecke wurden zunächst mittels einer subjektiven Umfragestudie die Einflussfaktoren auf eine anspruchsvolle Applikation bestimmt. Bei der Applikation handelt es sich um Cloud Gaming, bei dem die Berechnung der Videospiele Umgebung in der Cloud stattfindet und per Video über das Netz transportiert wird. Hier konnte gezeigt werden, dass neben dem wichtigsten QoS Einflussfaktor Paketverlust auf dem Downlink auch die Art des Spieltyps und dessen Geschwindigkeit eine Rolle spielen. Dies belegt, dass neben QoS auch der Applikationszustand wichtig ist und dem Netz mitgeteilt werden sollte. Da eine Umsetzung eines solchen zustandsbewussten SDNs für das Beispiel Cloud Gaming auf Grund dessen proprietärer Implementierung nicht möglich war, wurde in dieser Doktorarbeit auf die Applikation YouTube Video Streaming ausgewichen. Für diese sind mit Hilfe des Tools „YoMo“ Statusinformationen abfragbar und können zur Netzsteuerung genutzt werden. Es konnte gezeigt werden, dass eine SDN-basierte Realisierung eines applikationsbewussten Netzes deutliche Vorteile gegenüber klassischen Netzwerk Management Methoden aufweist und die Nutzerqualität trotz Störfaktoren erhalten werden kann. T3 - Würzburger Beiträge zur Leistungsbewertung Verteilter Systeme - 03/14 KW - Leistungsbewertung KW - Netzwerk KW - Software Defined Networking KW - Quality of Experience KW - Cloud Gaming KW - Performance Evaluation Y1 - 2014 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-100795 SN - 1432-8801 ER - TY - CHAP A1 - Schlosser, Daniel A1 - Jarschel, Michael A1 - Duelli, Michael A1 - Hoßfeld, Tobias A1 - Hoffmann, Klaus A1 - Hoffmann, Marco A1 - Morper, Hans Jochen A1 - Jurca, Dan A1 - Khan, Ashiq T1 - A Use Case Driven Approach to Network Virtualization N2 - In today's Internet, services are very different in their requirements on the underlying transport network. In the future, this diversity will increase and it will be more difficult to accommodate all services in a single network. A possible approach to cope with this diversity within future networks is the introduction of support for running isolated networks for different services on top of a single shared physical substrate. This would also enable easy network management and ensure an economically sound operation. End-customers will readily adopt this approach as it enables new and innovative services without being expensive. In order to arrive at a concept that enables this kind of network, it needs to be designed around and constantly checked against realistic use cases. In this contribution, we present three use cases for future networks. We describe functional blocks of a virtual network architecture, which are necessary to support these use cases within the network. Furthermore, we discuss the interfaces needed between the functional blocks and consider standardization issues that arise in order to achieve a global consistent control and management structure of virtual networks. KW - Virtualisierung KW - Datenkommunikationsnetz KW - Internet KW - Rechnernetz KW - Anwendungsfall KW - Netzvirtualisierung KW - Standardisierung KW - Use case KW - network virtualization KW - future Internet architecture KW - standardization Y1 - 2010 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-55611 N1 - Accepted at IEEE Kaleidoscope 2010 ER - TY - JOUR A1 - Atienza, Nieves A1 - de Castro, Natalia A1 - Cortés, Carmen A1 - Garrido, M. Ángeles A1 - Grima, Clara I. A1 - Hernández, Gregorio A1 - Márquez, Alberto A1 - Moreno-González, Auxiliadora A1 - Nöllenburg, Martin A1 - Portillo, José Ramón A1 - Reyes, Pedro A1 - Valenzuela, Jesús A1 - Trinidad Villar, Maria A1 - Wolff, Alexander T1 - Cover contact graphs N2 - We study problems that arise in the context of covering certain geometric objects called seeds (e.g., points or disks) by a set of other geometric objects called cover (e.g., a set of disks or homothetic triangles). We insist that the interiors of the seeds and the cover elements are pairwise disjoint, respectively, but they can touch. We call the contact graph of a cover a cover contact graph (CCG). We are interested in three types of tasks, both in the general case and in the special case of seeds on a line: (a) deciding whether a given seed set has a connected CCG, (b) deciding whether a given graph has a realization as a CCG on a given seed set, and (c) bounding the sizes of certain classes of CCG’s. Concerning (a) we give efficient algorithms for the case that seeds are points and show that the problem becomes hard if seeds and covers are disks. Concerning (b) we show that this problem is hard even for point seeds and disk covers (given a fixed correspondence between graph vertices and seeds). Concerning (c) we obtain upper and lower bounds on the number of CCG’s for point seeds. KW - Informatik Y1 - 2012 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-78845 ER - TY - THES A1 - Hock, David Rogér T1 - Analysis and Optimization of Resilient Routing in Core Communication Networks T1 - Analyse und Optimierung von ausfallsicherem Routing in Kernkommunikationsnetzen N2 - Routing is one of the most important issues in any communication network. It defines on which path packets are transmitted from the source of a connection to the destination. It allows to control the distribution of flows between different locations in the network and thereby is a means to influence the load distribution or to reach certain constraints imposed by particular applications. As failures in communication networks appear regularly and cannot be completely avoided, routing is required to be resilient against such outages, i.e., routing still has to be able to forward packets on backup paths even if primary paths are not working any more. Throughout the years, various routing technologies have been introduced that are very different in their control structure, in their way of working, and in their ability to handle certain failure cases. Each of the different routing approaches opens up their own specific questions regarding configuration, optimization, and inclusion of resilience issues. This monograph investigates, with the example of three particular routing technologies, some concrete issues regarding the analysis and optimization of resilience. It thereby contributes to a better general, technology-independent understanding of these approaches and of their diverse potential for the use in future network architectures. The first considered routing type, is decentralized intra-domain routing based on administrative IP link costs and the shortest path principle. Typical examples are common today's intra-domain routing protocols OSPF and IS-IS. This type of routing includes automatic restoration abilities in case of failures what makes it in general very robust even in the case of severe network outages including several failed components. Furthermore, special IP-Fast Reroute mechanisms allow for a faster reaction on outages. For routing based on link costs, traffic engineering, e.g. the optimization of the maximum relative link load in the network, can be done indirectly by changing the administrative link costs to adequate values. The second considered routing type, MPLS-based routing, is based on the a priori configuration of primary and backup paths, so-called Label Switched Paths. The routing layout of MPLS paths offers more freedom compared to IP-based routing as it is not restricted by any shortest path constraints but any paths can be setup. However, this in general involves a higher configuration effort. Finally, in the third considered routing type, typically centralized routing using a Software Defined Networking (SDN) architecture, simple switches only forward packets according to routing decisions made by centralized controller units. SDN-based routing layouts offer the same freedom as for explicit paths configured using MPLS. In case of a failure, new rules can be setup by the controllers to continue the routing in the reduced topology. However, new resilience issues arise caused by the centralized architecture. If controllers are not reachable anymore, the forwarding rules in the single nodes cannot be adapted anymore. This might render a rerouting in case of connection problems in severe failure scenarios infeasible. N2 - Routing stellt eine der zentralen Aufgaben in Kommunikationsnetzen dar. Es bestimmt darüber, auf welchem Weg Verkehr von der Quelle zum Ziel transportiert wird. Durch geschicktes Routing kann dadurch eine Verteilung der Verkehrsflüsse zum Beispiel zur Lastverteilung erreicht werden. Da Fehler in Kommunikationsnetzen nicht vollständig verhindert werden können, muss Routing insbesondere ausfallsicher sein, d.h., im Falle von Fehlern im Netz muss das Routing weiterhin in der Lage sein, Pakete auf alternativen Pfaden zum Ziel zu transportieren. Es existieren verschiedene gängige Routingverfahren und Technologien, die sich hinsichtlich Ihrer Arbeitsweise, Ihrer Kontrollstrukturen und Ihrer Funktionalität in bestimmten Fehlerszenarien unterscheiden. Für diese verschiedenen Ansätze ergeben sich jeweils eigene Fragestellungen hinsichtlich der Konfiguration, der Optimierung und der Berücksichtigung von Ausfallsicherheit. Diese Doktorarbeit behandelt am Beispiel bestimmter Technologien einige konkrete Fragestellungen zur Analyse und Optimierung der Ausfallsicherheit. Sie liefert damit einen Beitrag zum besseren generellen Verständnis verschiedenartiger Routingansätze und deren unterschiedlichen Potentials für den Einsatz in zukünftigen Netzarchitekturen. Zuerst wird dezentrales Routing behandelt, basierend auf administrativen Linkgewichten und dem Prinzip der kürzesten Pfade, wie es beispielsweise in den Protokollen IS-IS und OSPF genutzt wird. Diese Routingverfahren beinhalteten automatische Rekonvergenz-Mechanismen um im Falle von Fehlern auf der verbleibenden Netzstruktur weiterhin einen Transport von Verkehr zu ermöglichen. Spezielle IP-Fast Reroute Mechanismen ermöglichen zudem eine schnelle Reaktion im Falle von Fehlern. Routing basierend auf Linkgewichten lässt sich nur indirekt durch die Wahl geeigneter Gewichte beeinflussen und optimieren. Die zweite in der Doktorarbeit behandelte Routingart ist MPLS-basiertes Routing, bei dem Labels für Pakete verwendet werden und Pakete anhand sogenannter vorkonfigurierter Label Switched Paths weitergeleitet werden. Diese Technologie bietet mehr Freiheiten bei der Wahl des Pfadlayouts, was aber wiederum im Allgemeinen einen erhöhten Konfigurationsaufwand mit sich bringt. Schließlich greift die Doktorarbeit auch das Routing in SDN Netzen auf. Dort erfolgt eine Trennung von Control Plane und Data Plane, so dass einzelne dedizierte Controller die Routingentscheidungen festlegen und ansonsten einfache Switches mit reduzierter Komplexität den Verkehr lediglich entsprechend der festgelegten Regeln weiterleiten. Dies ermöglicht die größte Freiheit bei der Konfiguration des Routing bringt aber wiederum neue Fragestellungen bedingt durch die zentralen Kontrolleinheiten mit sich. T3 - Würzburger Beiträge zur Leistungsbewertung Verteilter Systeme - 02/14 KW - Leistungsbewertung KW - Verteiltes System KW - Routing KW - Netzwerk KW - Optimization KW - Routing KW - Software Defined Networking KW - Optimierung Y1 - 2014 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-101681 SN - 1432-8801 ER - TY - JOUR A1 - Wolff, Alexander A1 - Rutter, Iganz T1 - Augmenting the Connectivity of Planar and Geometric Graphs JF - Journal of Graph Algorithms and Applications N2 - In this paper we study connectivity augmentation problems. Given a connected graph G with some desirable property, we want to make G 2-vertex connected (or 2-edge connected) by adding edges such that the resulting graph keeps the property. The aim is to add as few edges as possible. The property that we consider is planarity, both in an abstract graph-theoretic and in a geometric setting, where vertices correspond to points in the plane and edges to straight-line segments. We show that it is NP-hard to � nd a minimum-cardinality augmentation that makes a planar graph 2-edge connected. For making a planar graph 2-vertex connected this was known. We further show that both problems are hard in the geometric setting, even when restricted to trees. The problems remain hard for higher degrees of connectivity. On the other hand we give polynomial-time algorithms for the special case of convex geometric graphs. We also study the following related problem. Given a planar (plane geometric) graph G, two vertices s and t of G, and an integer c, how many edges have to be added to G such that G is still planar (plane geometric) and contains c edge- (or vertex-) disjoint s{t paths? For the planar case we give a linear-time algorithm for c = 2. For the plane geometric case we give optimal worst-case bounds for c = 2; for c = 3 we characterize the cases that have a solution. Y1 - 2012 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-97587 ER - TY - THES A1 - Driewer, Frauke T1 - Teleoperation Interfaces in Human-Robot Teams T1 - Benutzerschnittstellen für Teleoperation in Mensch-Roboter Teams N2 - Diese Arbeit beschäftigt sich mit der Verbesserung von Mensch-Roboter Interaktion in Mensch-Roboter Teams für Teleoperation Szenarien, wie z.B. robotergestützte Feuerwehreinsätze. Hierbei wird ein Konzept und eine Architektur für ein System zur Unterstützung von Teleoperation von Mensch-Roboter Teams vorgestellt. Die Anforderungen an Informationsaustausch und -verarbeitung, insbesondere für die Anwendung Rettungseinsatz, werden ausgearbeitet. Weiterhin wird das Design der Benutzerschnittstellen für Mensch-Roboter Teams dargestellt und Prinzipien für Teleoperation-Systeme und Benutzerschnittstellen erarbeitet. Alle Studien und Ansätze werden in einem Prototypen-System implementiert und in verschiedenen Benutzertests abgesichert. Erweiterungsmöglichkeiten zum Einbinden von 3D Sensordaten und die Darstellung auf Stereovisualisierungssystemen werden gezeigt. N2 - This work deals with teams in teleoperation scenarios, where one human team partner (supervisor) guides and controls multiple remote entities (either robotic or human) and coordinates their tasks. Such a team needs an appropriate infrastructure for sharing information and commands. The robots need to have a level of autonomy, which matches the assigned task. The humans in the team have to be provided with autonomous support, e.g. for information integration. Design and capabilities of the human-robot interfaces will strongly influence the performance of the team as well as the subjective feeling of the human team partners. Here, it is important to elaborate the information demand as well as how information is presented. Such human-robot systems need to allow the supervisor to gain an understanding of what is going on in the remote environment (situation awareness) by providing the necessary information. This includes achieving fast assessment of the robot´s or remote human´s state. Processing, integration and organization of data as well as suitable autonomous functions support decision making and task allocation and help to decrease the workload in this multi-entity teleoperation task. Interaction between humans and robots is improved by a common world model and a responsive system and robots. The remote human profits from a simplified user interface providing exactly the information needed for the actual task at hand. The topic of this thesis is the investigation of such teleoperation interfaces in human-robot teams, especially for high-risk, time-critical, and dangerous tasks. The aim is to provide a suitable human-robot team structure as well as analyze the demands on the user interfaces. On one side, it will be looked on the theoretical background (model, interactions, and information demand). On the other side, real implementations for system, robots, and user interfaces are presented and evaluated as testbeds for the claimed requirements. Rescue operations, more precisely fire-fighting, was chosen as an exemplary application scenario for this work. The challenges in such scenarios are high (highly dynamic environments, high risk, time criticality etc.) and it can be expected that results can be transferred to other applications, which have less strict requirements. The present work contributes to the introduction of human-robot teams in task-oriented scenarios, such as working in high risk domains, e.g. fire-fighting. It covers the theoretical background of the required system, the analysis of related human factors concepts, as well as discussions on implementation. An emphasis is placed on user interfaces, their design, requirements and user testing, as well as on the used techniques (three-dimensional sensor data representation, mixed reality, and user interface design guidelines). Further, the potential integration of 3D sensor data as well as the visualization on stereo visualization systems is introduced. T3 - Forschungsberichte in der Robotik = Research Notes in Robotics - 1 KW - Robotik KW - Mobiler Roboter KW - Autonomer Roboter KW - Mensch-Maschine-System KW - Mensch-Maschine-Schnittstelle KW - Mixed Reality KW - Mensch-Roboter-Interaktion KW - Situationsbewusstsein KW - Teleoperation KW - Benutzerschnittstelle KW - Rettungsroboter KW - Human-Robot-Interaction KW - Situation Awareness KW - Teleoperation KW - User Interface KW - Search-and-Rescue Y1 - 2008 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-36351 SN - 978-3-923959-57-0 N1 - Zugl. gedruckte Ausgabe: ISBN: 978-3-923959-54-9 (Paper) ER - TY - THES A1 - Hess, Martin T1 - Motion coordination and control in systems of nonholonomic autonomous vehicles T1 - Bewegungskoordination und Regelung in Gruppen nichtholonomer autonomer Fahrzeuge N2 - This work focuses on coordination methods and the control of motion in groups of nonholonomic wheeled mobile robots, in particular of the car-like type. These kind of vehicles are particularly restricted in their mobility. In the main part of this work the two problems of formation motion coordination and of rendezvous in distributed multi-vehicle systems are considered. We introduce several enhancements to an existing motion planning approach for formations of nonholonomic mobile robots. Compared to the original method, the extended approach is able to handle time-varying reference speeds as well as adjustments of the formation's shape during reference trajectory segments with continuously differentiable curvature. Additionally, undesired discontinuities in the speed and steering profiles of the vehicles are avoided. Further, the scenario of snow shoveling on an airfield by utilizing multiple formations of autonomous snowplows is discussed. We propose solutions to the subproblems of motion planning for the formations and tracking control for the individual vehicles. While all situations that might occur have been tested in a simulation environment, we also verified the developed tracking controller in real robot hardware experiments. The task of the rendezvous problem in groups of car-like robots is to drive all vehicles to a common position by means of decentralized control laws. Typically there exists no direct interaction link between all of the vehicles. In this work we present decentralized rendezvous control laws for vehicles with free and with bounded steering. The convergence properties of the approaches are analyzed by utilizing Lyapunov based techniques. Furthermore, they are evaluated within various simulation experiments, while the bounded steering case is also verified within laboratory hardware experiments. Finally we introduce a modification to the bounded steering system that increases the convergence speed at the expense of a higher traveled distance of the vehicles. N2 - Diese Arbeit befasst sich mit Methoden zur Bewegungskoordination und Regelung in Gruppen autonomer, nichtholomer Fahrzeuge, wobei vornehmlich autoähnliche mobile Roboter betrachtet werden. Diese Fahrzeuge sind besonders eingeschränkt in ihrer Bewegungsfreiheit. Im Hauptteil der Arbeit werden die Probleme der Formationsbewegung und des Rendezvous in Gruppen verteilter Fahrzeuge betrachtet. Für ein bestehendes Verfahren zur Bewegungsplanung für Formationen nichtholonomer, mobiler Fahrzeuge werden eine Reihe von Verbesserungen vorgestellt. Diese ermöglichen dem erweiterten Verfahren den Umgang mit Referenztrajektorien mit nicht-konstanter Geschwindigkeit und stetig differenzierbarer Krümmung. Außerdem werden im Gegensatz zum ursprünglichen Ansatz unerwünschte Sprungstellen in den Geschwindigkeits- und Krümmungsprofilen der Fahrzeuge vermieden. Desweiteren wird in dieser Arbeit das Schneeräumen auf Flughafenrollfeldern mittels Formationen autonomer Schneepflugfahrzeuge diskutiert. Dabei werden Lösungen für die Teilprobleme Bewegungsplanung der Formationen und Spurführungsregelung der einzelnen Fahrzeuge vorgestellt. Zusätzlich zu den durchgeführten Simulationen werden außerdem die Ergebnisse von Hardwareexperimenten mit der entwickelten Spurführungsregelung präsentiert. Das Rendezvous-Problem in Gruppen autoähnlicher Roboter besteht darin, die Fahrzeuge durch dezentrale Regelgesetze zu einer gemeinsamen Position in der Ebene zu bewegen. Dabei besteht typischerweise keine direkte Interaktionsverbindung zwischen allen Fahrzeugen der Gruppe. In dieser Arbeit werden verteilte Rendezvous-Regelgesetze für Fahrzeuge mit unbeschränktem und mit beschränktem Lenkwinkel eingeführt und deren Konvergenzeigenschaften mit Methoden zur Stabilitätsanalyse nicht-linearer Systeme untersucht. Die vorgestellten Regelsysteme werden weiterhin anhand von Simulationen und Hardwareexperimenten verifiziert. Schließlich wird noch eine Erweiterung des Systems für Fahrzeuge mit beschränktem Lenkwinkel vorgestellt, welches die Konvergenzgeschwindigkeit auf Kosten der zurückgelegten Distanz der Fahrzeuge erhöht. T3 - Forschungsberichte in der Robotik = Research Notes in Robotics - 2 KW - Robotik KW - Autonomer Roboter KW - Mobiler Roboter KW - Bewegungskoordination KW - Dezentrale Regelung KW - Nichtlineare Regelung KW - Rendezvous KW - Kooperierende mobile Roboter KW - Formationsbewegung KW - Nichtholonome Fahrzeuge KW - Mehrfahrzeugsysteme KW - Unstetige Regelung KW - Autonomous multi-vehicle systems KW - multi-vehicle formations KW - multi-vehicle rendezvous KW - car-like robots KW - nonholonomic vehicles Y1 - 2009 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-46442 SN - 978-3-923959-55-6 ER - TY - THES A1 - Kaempgen, Benedikt T1 - Deskriptives Data-Mining für Entscheidungsträger: Eine Mehrfachfallstudie T1 - Descriptive data mining for decision-makers: a multiple case study N2 - Das Potenzial der Wissensentdeckung in Daten wird häufig nicht ausgenutzt, was hauptsächlich auf Barrieren zwischen dem Entwicklerteam und dem Endnutzer des Data-Mining zurückzuführen ist. In dieser Arbeit wird ein transparenter Ansatz zum Beschreiben und Erklären von Daten für Entscheidungsträger vorgestellt. In Entscheidungsträger-zentrierten Aufgaben werden die Projektanforderungen definiert und die Ergebnisse zu einer Geschichte zusammengestellt. Eine Anforderung besteht dabei aus einem tabellarischen Bericht und ggf. Mustern in seinem Inhalt, jeweils verständlich für einen Entscheidungsträger. Die technischen Aufgaben bestehen aus einer Datenprüfung, der Integration der Daten in einem Data-Warehouse sowie dem Generieren von Berichten und dem Entdecken von Mustern wie in den Anforderungen beschrieben. Mehrere Data-Mining-Projekte können durch Wissensmanagement sowie eine geeignete Infrastruktur voneinander profitieren. Der Ansatz wurde in zwei Projekten unter Verwendung von ausschließlich Open-Source-Software angewendet. N2 - Despite high potential of data mining in business and science many projects fail due to barriers between the developer team and the end user. In this work a more transparent approach to describing and explaining data to a decision-maker is presented. In decision-maker-centric tasks project requirements are defined and finally the results composed to a story. A requirement is made of a tabular report and possibly patterns in its data, each understandable to a decision-maker. The technical tasks consist of a data assay, the integration of data within a data warehouse and, as required, the creation of reports and the discovery of patterns. Multiple data mining projects benefit from each other through knowledge management and a common infrastructure. The approach has been applied to two projects exclusively using open source systems. KW - Data Mining KW - Entscheidungsträger KW - Fallstudie KW - Methodologie KW - Endnutzer KW - Business Intelligence KW - Open Source KW - data mining KW - case study KW - process model KW - end user KW - open source Y1 - 2009 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-46343 ER - TY - JOUR A1 - Schlör, Daniel A1 - Ring, Markus A1 - Hotho, Andreas T1 - iNALU: Improved Neural Arithmetic Logic Unit JF - Frontiers in Artificial Intelligence N2 - Neural networks have to capture mathematical relationships in order to learn various tasks. They approximate these relations implicitly and therefore often do not generalize well. The recently proposed Neural Arithmetic Logic Unit (NALU) is a novel neural architecture which is able to explicitly represent the mathematical relationships by the units of the network to learn operations such as summation, subtraction or multiplication. Although NALUs have been shown to perform well on various downstream tasks, an in-depth analysis reveals practical shortcomings by design, such as the inability to multiply or divide negative input values or training stability issues for deeper networks. We address these issues and propose an improved model architecture. We evaluate our model empirically in various settings from learning basic arithmetic operations to more complex functions. Our experiments indicate that our model solves stability issues and outperforms the original NALU model in means of arithmetic precision and convergence. KW - neural networks KW - machine learning KW - arithmetic calculations KW - neural architecture KW - experimental evaluation Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-212301 SN - 2624-8212 VL - 3 ER - TY - RPRT A1 - Vomhoff, Viktoria A1 - Geissler, Stefan A1 - Gebert, Steffen A1 - Hossfeld, Tobias T1 - Towards Understanding the Global IPX Network from an MVNO Perspective T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - In this paper, we work to understand the global IPX network from the perspective of an MVNO. In order to do this, we provide a brief description of the global architecture of mobile carriers. We provide initial results with respect to mapping the vast and complex interconnection network enabling global roaming from the point of view of a single MVNO. Finally, we provide preliminary results regarding the quality of service observed under global roaming conditions. KW - global IPX network Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322121 ER - TY - JOUR A1 - Loh, Frank A1 - Poignée, Fabian A1 - Wamser, Florian A1 - Leidinger, Ferdinand A1 - Hoßfeld, Tobias T1 - Uplink vs. Downlink: Machine Learning-Based Quality Prediction for HTTP Adaptive Video Streaming JF - Sensors N2 - Streaming video is responsible for the bulk of Internet traffic these days. For this reason, Internet providers and network operators try to make predictions and assessments about the streaming quality for an end user. Current monitoring solutions are based on a variety of different machine learning approaches. The challenge for providers and operators nowadays is that existing approaches require large amounts of data. In this work, the most relevant quality of experience metrics, i.e., the initial playback delay, the video streaming quality, video quality changes, and video rebuffering events, are examined using a voluminous data set of more than 13,000 YouTube video streaming runs that were collected with the native YouTube mobile app. Three Machine Learning models are developed and compared to estimate playback behavior based on uplink request information. The main focus has been on developing a lightweight approach using as few features and as little data as possible, while maintaining state-of-the-art performance. KW - HTTP adaptive video streaming KW - quality of experience prediction KW - machine learning Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-241121 SN - 1424-8220 VL - 21 IS - 12 ER - TY - CHAP A1 - Truman, Samuel A1 - von Mammen, Sebastian T1 - Interactive Self-Assembling Agent Ensembles T2 - Proceedings of the 1st Games Technology Summit N2 - In this paper, we bridge the gap between procedural content generation (PCG) and user-generated content (UGC) by proposing and demonstrating an interactive agent-based model of self-assembling ensembles that can be directed though user input. We motivate these efforts by considering the opportunities technology provides to pursue game designs based on according game design frameworks. We present three different use cases of the proposed model that emphasize its potential to (1) self-assemble into predefined 3D graphical assets, (2) define new structures in the context of virtual environments by self-assembling layers on the surfaces of arbitrary 3D objects, and (3) allow novel structures to self-assemble only considering the model’s configuration and no external dependencies. To address the performance restrictions in computer games, we realized the prototypical model implementation by means of an efficient entity component system (ECS). We conclude the paper with an outlook on future steps to further explore novel interactive, dynamic PCG mechanics and to ensure their efficiency. KW - procedural content generation KW - user-generated content KW - game mechanics KW - agent-based models KW - self-assembly Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-246032 ER - TY - RPRT A1 - Navade, Piyush A1 - Maile, Lisa A1 - German, Reinhard T1 - Multiple DCLC Routing Algorithms for Ultra-Reliable and Time-Sensitive Applications T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - This paper discusses the problem of finding multiple shortest disjoint paths in modern communication networks, which is essential for ultra-reliable and time-sensitive applications. Dijkstra’s algorithm has been a popular solution for the shortest path problem, but repetitive use of it to find multiple paths is not scalable. The Multiple Disjoint Path Algorithm (MDPAlg), published in 2021, proposes the use of a single full graph to construct multiple disjoint paths. This paper proposes modifications to the algorithm to include a delay constraint, which is important in time-sensitive applications. Different delay constraint least-cost routing algorithms are compared in a comprehensive manner to evaluate the benefits of the adapted MDPAlg algorithm. Fault tolerance, and thereby reliability, is ensured by generating multiple link-disjoint paths from source to destination. KW - Dijkstra’s algorithm KW - shortest path routing KW - disjoint multi-paths KW - delay constrained KW - least cost Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322177 ER - TY - JOUR A1 - Halbig, Andreas A1 - Latoschik, Marc Erich T1 - A systematic review of physiological measurements, factors, methods, and applications in virtual reality JF - Frontiers in Virtual Reality N2 - Measurements of physiological parameters provide an objective, often non-intrusive, and (at least semi-)automatic evaluation and utilization of user behavior. In addition, specific hardware devices of Virtual Reality (VR) often ship with built-in sensors, i.e. eye-tracking and movements sensors. Hence, the combination of physiological measurements and VR applications seems promising. Several approaches have investigated the applicability and benefits of this combination for various fields of applications. However, the range of possible application fields, coupled with potentially useful and beneficial physiological parameters, types of sensor, target variables and factors, and analysis approaches and techniques is manifold. This article provides a systematic overview and an extensive state-of-the-art review of the usage of physiological measurements in VR. We identified 1,119 works that make use of physiological measurements in VR. Within these, we identified 32 approaches that focus on the classification of characteristics of experience, common in VR applications. The first part of this review categorizes the 1,119 works by field of application, i.e. therapy, training, entertainment, and communication and interaction, as well as by the specific target factors and variables measured by the physiological parameters. An additional category summarizes general VR approaches applicable to all specific fields of application since they target typical VR qualities. In the second part of this review, we analyze the target factors and variables regarding the respective methods used for an automatic analysis and, potentially, classification. For example, we highlight which measurement setups have been proven to be sensitive enough to distinguish different levels of arousal, valence, anxiety, stress, or cognitive workload in the virtual realm. This work may prove useful for all researchers wanting to use physiological data in VR and who want to have a good overview of prior approaches taken, their benefits and potential drawbacks. KW - virtual reality KW - use cases KW - sesnsors KW - tools KW - biosignals KW - psychophyisology KW - HMD (Head-Mounted Display) KW - systematic review Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-260503 VL - 2 ER - TY - JOUR A1 - Li, Ningbo A1 - Guan, Lianwu A1 - Gao, Yanbin A1 - Du, Shitong A1 - Wu, Menghao A1 - Guang, Xingxing A1 - Cong, Xiaodan T1 - Indoor and outdoor low-cost seamless integrated navigation system based on the integration of INS/GNSS/LIDAR system JF - Remote Sensing N2 - Global Navigation Satellite System (GNSS) provides accurate positioning data for vehicular navigation in open outdoor environment. In an indoor environment, Light Detection and Ranging (LIDAR) Simultaneous Localization and Mapping (SLAM) establishes a two-dimensional map and provides positioning data. However, LIDAR can only provide relative positioning data and it cannot directly provide the latitude and longitude of the current position. As a consequence, GNSS/Inertial Navigation System (INS) integrated navigation could be employed in outdoors, while the indoors part makes use of INS/LIDAR integrated navigation and the corresponding switching navigation will make the indoor and outdoor positioning consistent. In addition, when the vehicle enters the garage, the GNSS signal will be blurred for a while and then disappeared. Ambiguous GNSS satellite signals will lead to the continuous distortion or overall drift of the positioning trajectory in the indoor condition. Therefore, an INS/LIDAR seamless integrated navigation algorithm and a switching algorithm based on vehicle navigation system are designed. According to the experimental data, the positioning accuracy of the INS/LIDAR navigation algorithm in the simulated environmental experiment is 50% higher than that of the Dead Reckoning (DR) algorithm. Besides, the switching algorithm developed based on the INS/LIDAR integrated navigation algorithm can achieve 80% success rate in navigation mode switching. KW - vehicular navigation KW - GNSS/INS integrated navigation KW - INS/LIDAR integrated navigation KW - switching navigation Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-216229 SN - 2072-4292 VL - 12 IS - 19 ER - TY - JOUR A1 - Lesch, Veronika A1 - König, Maximilian A1 - Kounev, Samuel A1 - Stein, Anthony A1 - Krupitzer, Christian T1 - Tackling the rich vehicle routing problem with nature-inspired algorithms JF - Applied Intelligence N2 - In the last decades, the classical Vehicle Routing Problem (VRP), i.e., assigning a set of orders to vehicles and planning their routes has been intensively researched. As only the assignment of order to vehicles and their routes is already an NP-complete problem, the application of these algorithms in practice often fails to take into account the constraints and restrictions that apply in real-world applications, the so called rich VRP (rVRP) and are limited to single aspects. In this work, we incorporate the main relevant real-world constraints and requirements. We propose a two-stage strategy and a Timeline algorithm for time windows and pause times, and apply a Genetic Algorithm (GA) and Ant Colony Optimization (ACO) individually to the problem to find optimal solutions. Our evaluation of eight different problem instances against four state-of-the-art algorithms shows that our approach handles all given constraints in a reasonable time. KW - logistics KW - rich vehicle routing problem KW - ant-colony optimization KW - genetic algorithm KW - real-world application Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-268942 SN - 1573-7497 VL - 52 ER - TY - RPRT A1 - Deutschmann, Jörg A1 - Hielscher, Kai-Steffen A1 - German, Reinhard T1 - Next-Generation Satellite Communication Networks T2 - Würzburg Workshop on Next-Generation Communication Networks (WueWoWas'22) N2 - This paper gives an overview of our recent activities in the field of satellite communication networks, including an introduction to geostationary satellite systems and Low Earth Orbit megaconstellations. To mitigate the high latencies of geostationary satellite networks, TCP-splitting Performance Enhancing Proxies are deployed. However, these cannot be applied in the case of encrypted transport headers as it is the case for VPNs or QUIC. We summarize performance evaluation results from multiple measurement campaigns. In a recently concluded project, multipath communication was used to combine the advantages of very heterogeneous communication paths: low data rate, low latency (e.g., DSL light) and high data rate, high latency (e.g., geostationary satellite). KW - Datennetz KW - satellite communication KW - Performance Enhancing Proxies KW - transport protocols KW - VPN KW - QUIC KW - multipath communication KW - hybrid access Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-280732 ER - TY - JOUR A1 - Carolus, Astrid A1 - Wienrich, Carolin A1 - Törke, Anna A1 - Friedel, Tobias A1 - Schwietering, Christian A1 - Sperzel, Mareike T1 - ‘Alexa, I feel for you!’ Observers’ empathetic reactions towards a conversational agent JF - Frontiers in Computer Science N2 - Conversational agents and smart speakers have grown in popularity offering a variety of options for use, which are available through intuitive speech operation. In contrast to the standard dyad of a single user and a device, voice-controlled operations can be observed by further attendees resulting in new, more social usage scenarios. Referring to the concept of ‘media equation’ and to research on the idea of ‘computers as social actors,’ which describes the potential of technology to trigger emotional reactions in users, this paper asks for the capacity of smart speakers to elicit empathy in observers of interactions. In a 2 × 2 online experiment, 140 participants watched a video of a man talking to an Amazon Echo either rudely or neutrally (factor 1), addressing it as ‘Alexa’ or ‘Computer’ (factor 2). Controlling for participants’ trait empathy, the rude treatment results in participants’ significantly higher ratings of empathy with the device, compared to the neutral treatment. The form of address had no significant effect. Results were independent of the participants’ gender and usage experience indicating a rather universal effect, which confirms the basic idea of the media equation. Implications for users, developers and researchers were discussed in the light of (future) omnipresent voice-based technology interaction scenarios. KW - conversational agent KW - empathy KW - smart speaker KW - media equation KW - computers as social actors KW - human-computer interaction Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-258807 VL - 3 ER - TY - RPRT A1 - Riegler, Clemens A1 - Werner, Lennart A1 - Kayal, Hakan T1 - MAPLE: Marsian Autorotation Probe Lander Experiment N2 - The first step towards aerial planetary exploration has been made. Ingenuity shows extremely promising results, and new missions are already underway. Rotorcraft are capable of flight. This capability could be utilized to support the last stages of Entry, Descent, and Landing. Thus, mass and complexity could be scaled down. Autorotation is one method of descent. It describes unpowered descent and landing, typically performed by helicopters in case of an engine failure. MAPLE is suggested to test these procedures and understand autorotation on other planets. In this series of experiments, the Ingenuity helicopter is utilized. Ingenuity would autorotate a ”mid-air-landing” before continuing with normal flight. Ultimately, the collected data shall help to understand autorotation on Mars and its utilization for interplanetary exploration. T3 - Raumfahrttechnik und Extraterrestrik - 2 KW - autorotation KW - descent KW - Mars KW - rotorcraft KW - landing KW - aerospace Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-282390 ER - TY - JOUR A1 - Obremski, David A1 - Lugrin, Jean-Luc A1 - Schaper, Philipp A1 - Lugrin, Birgit T1 - Non-native speaker perception of Intelligent Virtual Agents in two languages: the impact of amount and type of grammatical mistakes JF - Journal on Multimodal User Interfaces N2 - Having a mixed-cultural membership becomes increasingly common in our modern society. It is thus beneficial in several ways to create Intelligent Virtual Agents (IVAs) that reflect a mixed-cultural background as well, e.g., for educational settings. For research with such IVAs, it is essential that they are classified as non-native by members of a target culture. In this paper, we focus on variations of IVAs’ speech to create the impression of non-native speakers that are identified as such by speakers of two different mother tongues. In particular, we investigate grammatical mistakes and identify thresholds beyond which the agents is clearly categorised as a non-native speaker. Therefore, we conducted two experiments: one for native speakers of German, and one for native speakers of English. Results of the German study indicate that beyond 10% of word order mistakes and 25% of infinitive mistakes German-speaking IVAs are perceived as non-native speakers. Results of the English study indicate that beyond 50% of omission mistakes and 50% of infinitive mistakes English-speaking IVAs are perceived as non-native speakers. We believe these thresholds constitute helpful guidelines for computational approaches of non-native speaker generation, simplifying research with IVAs in mixed-cultural settings. KW - mixed-cultural settings KW - Intelligent Virtual Agents KW - verbal behaviour Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-269984 SN - 1783-8738 VL - 15 IS - 2 ER - TY - JOUR A1 - Reinhard, Sebastian A1 - Helmerich, Dominic A. A1 - Boras, Dominik A1 - Sauer, Markus A1 - Kollmannsberger, Philip T1 - ReCSAI: recursive compressed sensing artificial intelligence for confocal lifetime localization microscopy JF - BMC Bioinformatics N2 - Background Localization-based super-resolution microscopy resolves macromolecular structures down to a few nanometers by computationally reconstructing fluorescent emitter coordinates from diffraction-limited spots. The most commonly used algorithms are based on fitting parametric models of the point spread function (PSF) to a measured photon distribution. These algorithms make assumptions about the symmetry of the PSF and thus, do not work well with irregular, non-linear PSFs that occur for example in confocal lifetime imaging, where a laser is scanned across the sample. An alternative method for reconstructing sparse emitter sets from noisy, diffraction-limited images is compressed sensing, but due to its high computational cost it has not yet been widely adopted. Deep neural network fitters have recently emerged as a new competitive method for localization microscopy. They can learn to fit arbitrary PSFs, but require extensive simulated training data and do not generalize well. A method to efficiently fit the irregular PSFs from confocal lifetime localization microscopy combining the advantages of deep learning and compressed sensing would greatly improve the acquisition speed and throughput of this method. Results Here we introduce ReCSAI, a compressed sensing neural network to reconstruct localizations for confocal dSTORM, together with a simulation tool to generate training data. We implemented and compared different artificial network architectures, aiming to combine the advantages of compressed sensing and deep learning. We found that a U-Net with a recursive structure inspired by iterative compressed sensing showed the best results on realistic simulated datasets with noise, as well as on real experimentally measured confocal lifetime scanning data. Adding a trainable wavelet denoising layer as prior step further improved the reconstruction quality. Conclusions Our deep learning approach can reach a similar reconstruction accuracy for confocal dSTORM as frame binning with traditional fitting without requiring the acquisition of multiple frames. In addition, our work offers generic insights on the reconstruction of sparse measurements from noisy experimental data by combining compressed sensing and deep learning. We provide the trained networks, the code for network training and inference as well as the simulation tool as python code and Jupyter notebooks for easy reproducibility. KW - compressed sensing KW - AI KW - SMLM KW - FLIMbee KW - dSTORM Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-299768 VL - 23 IS - 1 ER - TY - RPRT A1 - Simon, Manuel A1 - Gallenmüller, Sebastian A1 - Carle, Georg T1 - Never Miss Twice - Add-On-Miss Table Updates in Software Data Planes T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - State Management at line rate is crucial for critical applications in next-generation networks. P4 is a language used in software-defined networking to program the data plane. The data plane can profit in many circumstances when it is allowed to manage its state without any detour over a controller. This work is based on a previous study by investigating the potential and performance of add-on-miss insertions of state by the data plane. The state keeping capabilities of P4 are limited regarding the amount of data and the update frequency. We follow the tentative specification of an upcoming portable-NIC-architecture and implement these changes into the software P4 target T4P4S. We show that insertions are possible with only a slight overhead compared to lookups and evaluate the influence of the rate of insertions on their latency. KW - SDN KW - state management KW - P4 KW - Add-on-Miss Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322071 ER - TY - RPRT A1 - Brisch, Fabian A1 - Kassler, Andreas A1 - Vestin, Jonathan A1 - Pieska, Marcus A1 - Amend, Markus T1 - Accelerating Transport Layer Multipath Packet Scheduling for 5G-ATSSS T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - Utilizing multiple access networks such as 5G, 4G, and Wi-Fi simultaneously can lead to increased robustness, resiliency, and capacity for mobile users. However, transparently implementing packet distribution over multiple paths within the core of the network faces multiple challenges including scalability to a large number of customers, low latency, and high-capacity packet processing requirements. In this paper, we offload congestion-aware multipath packet scheduling to a smartNIC. However, such hardware acceleration faces multiple challenges due to programming language and platform limitations. We implement different multipath schedulers in P4 with different complexity in order to cope with dynamically changing path capacities. Using testbed measurements, we show that our CMon scheduler, which monitors path congestion in the data plane and dynamically adjusts scheduling weights for the different paths based on path state information, can process more than 3.5 Mpps packets 25 μs latency. KW - multipath packet scheduling KW - P4 KW - MP-DCCP KW - 5G KW - ATSSSS Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322052 ER - TY - RPRT A1 - Hasslinger, Gerhard A1 - Ntougias, Konstantinos A1 - Hasslinger, Frank A1 - Hohlfeld, Oliver T1 - Performance Analysis of Basic Web Caching Strategies (LFU, LRU, FIFO, ...) with Time-To-Live Data Validation T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - Web caches often use a Time-to-live (TTL) limit to validate data consistency with web servers. We study the impact of TTL constraints on the hit ratio of basic strategies in caches of fixed size. We derive analytical results and confirm their accuracy in comparison to simulations. We propose a score-based caching method with awareness of the current TTL per data for improving the hit ratio close to the upper bound. KW - LRU KW - LFU KW - FIFO caching strategies KW - hit ratio analysis and simulation KW - TTL validation of data consistency Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322048 ER - TY - RPRT A1 - Funda, Christoph A1 - Marín García, Pablo A1 - German, Reinhard A1 - Hielscher, Kai-Steffen T1 - Online Algorithm for Arrival & Service Curve Estimation T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - This paper presents a novel concept to extend state-of-the-art buffer monitoring with additional measures to estimate service-curves. The online algorithm for service-curve estimation replaces the state-of-the-art timestamp logging, as we expect it to overcome the main disadvantages of generating a huge amount of data and using a lot of CPU resources to store the data to a file during operation. We prove the accuracy of the online-algorithm offline with timestamp data and compare the derived bounds to the measured delay and backlog. We also do a proof-of- concept of the online-algorithm, implement it in LabVIEW and compare its performance to the timestamp logging by CPU load and data-size of the log-file. However, the implementation is still work-in-progress. KW - hardware-in-the-loop streaming system KW - network calculus KW - service-curve estimation KW - performance monitoring Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322112 ER - TY - RPRT A1 - Mazigh, Sadok Mehdi A1 - Beausencourt, Marcel A1 - Bode, Max Julius A1 - Scheffler, Thomas T1 - Using P4-INT on Tofino for Measuring Device Performance Characteristics in a Network Lab T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - This paper presents a prototypical implementation of the In-band Network Telemetry (INT) specification in P4 and demonstrates a use case, where a Tofino Switch is used to measure device and network performance in a lab setting. This work is based on research activities in the area of P4 data plane programming conducted at the network lab of HTW Berlin. KW - P4-INT Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322084 ER - TY - RPRT A1 - Nguyen, Kien A1 - Loh, Frank A1 - Hoßfeld, Tobias T1 - Challenges of Serverless Deployment in Edge-MEC-Cloud T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - The emerging serverless computing may meet Edge Cloud in a beneficial manner as the two offer flexibility and dynamicity in optimizing finite hardware resources. However, the lack of proper study of a joint platform leaves a gap in literature about consumption and performance of such integration. To this end, this paper identifies the key questions and proposes a methodology to answer them. KW - Edge-MEC-Cloud Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322025 ER - TY - RPRT A1 - Raffeck, Simon A1 - Geißler, Stefan A1 - Hoßfeld, Tobias T1 - Towards Understanding the Signaling Traffic in 5G Core Networks T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - The Fifth Generation (5G) communication technology, its infrastructure and architecture, though already deployed in campus and small scale networks, is still undergoing continuous changes and research. Especially, in the light of future large scale deployments and industrial use cases, a detailed analysis of the performance and utilization with regard to latency and service times constraints is crucial. To this end, a fine granular investigation of the Network Function (NF) based core system and the duration for all the tasks performed by these services is necessary. This work presents the first steps towards analyzing the signaling traffic in 5G core networks, and introduces a tool to automatically extract sequence diagrams and service times for NF tasks from traffic traces. KW - signaling traffic KW - 5G core network Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322106 ER - TY - RPRT A1 - Großmann, Marcel A1 - Homeyer, Tobias T1 - Emulation of Multipath Transmissions in P4 Networks with Kathará T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - Packets sent over a network can either get lost or reach their destination. Protocols like TCP try to solve this problem by resending the lost packets. However, retransmissions consume a lot of time and are cumbersome for the transmission of critical data. Multipath solutions are quite common to address this reliability issue and are available on almost every layer of the ISO/OSI model. We propose a solution based on a P4 network to duplicate packets in order to send them to their destination via multiple routes. The last network hop ensures that only a single copy of the traffic is further forwarded to its destination by adopting a concept similar to Bloom filters. Besides, if fast delivery is requested we provide a P4 prototype, which randomly forwards the packets over different transmission paths. For reproducibility, we implement our approach in a container-based network emulation system called Kathará. KW - P4 KW - multipath KW - emulation KW - Kathará Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322095 ER - TY - RPRT A1 - Riegler, Clemens A1 - Kayal, Hakan T1 - VELEX: Venus Lightning Experiment N2 - Lightning has fascinated humanity since the beginning of our existence. Different types of lightning like sprites and blue jets were discovered, and many more are theorized. However, it is very likely that these phenomena are not exclusive to our home planet. Venus’s dense and active atmosphere is a place where lightning is to be expected. Missions like Venera, Pioneer, and Galileo have carried instruments to measure electromagnetic activity. These measurements have indeed delivered results. However, these results are not clear. They could be explained by other effects like cosmic rays, plasma noise, or spacecraft noise. Furthermore, these lightning seem different from those we know from our home planet. In order to tackle these issues, a different approach to measurement is proposed. When multiple devices in different spacecraft or locations can measure the same atmospheric discharge, most other explanations become increasingly less likely. Thus, the suggested instrument and method of VELEX incorporates multiple spacecraft. With this approach, the question about the existence of lightning on Venus could be settled. T3 - Raumfahrttechnik und Extraterrestrik - 3 KW - Venus KW - Lightning KW - CubeSat KW - Balloon KW - Autorotation Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-282481 ER - TY - JOUR A1 - Wamser, Florian A1 - Seufert, Anika A1 - Hall, Andrew A1 - Wunderer, Stefan A1 - Hoßfeld, Tobias T1 - Valid statements by the crowd: statistical measures for precision in crowdsourced mobile measurements JF - Network N2 - Crowdsourced network measurements (CNMs) are becoming increasingly popular as they assess the performance of a mobile network from the end user's perspective on a large scale. Here, network measurements are performed directly on the end-users' devices, thus taking advantage of the real-world conditions end-users encounter. However, this type of uncontrolled measurement raises questions about its validity and reliability. The problem lies in the nature of this type of data collection. In CNMs, mobile network subscribers are involved to a large extent in the measurement process, and collect data themselves for the operator. The collection of data on user devices in arbitrary locations and at uncontrolled times requires means to ensure validity and reliability. To address this issue, our paper defines concepts and guidelines for analyzing the precision of CNMs; specifically, the number of measurements required to make valid statements. In addition to the formal definition of the aspect, we illustrate the problem and use an extensive sample data set to show possible assessment approaches. This data set consists of more than 20.4 million crowdsourced mobile measurements from across France, measured by a commercial data provider. KW - mobile networks KW - crowdsourced measurements KW - statistical validity Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-284154 SN - 2673-8732 VL - 1 IS - 2 SP - 215 EP - 232 ER - TY - RPRT A1 - Grigorjew, Alexej A1 - Schumann, Lukas Kilian A1 - Diederich, Philip A1 - Hoßfeld, Tobias A1 - Kellerer, Wolfgang T1 - Understanding the Performance of Different Packet Reception and Timestamping Methods in Linux T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - This document briefly presents some renowned packet reception techniques for network packets in Linux systems. Further, it compares their performance when measuring packet timestamps with respect to throughput and accuracy. Both software and hardware timestamps are compared, and various parameters are examined, including frame size, link speed, network interface card, and CPU load. The results indicate that hardware timestamping offers significantly better accuracy with no downsides, and that packet reception techniques that avoid system calls offer superior measurement throughput. KW - packet reception method KW - timestamping method KW - Linux Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322064 ER - TY - THES A1 - Somody, Joseph Christian Campbell T1 - Leveraging deep learning for identification and structural determination of novel protein complexes from \(in\) \(situ\) electron cryotomography of \(Mycoplasma\) \(pneumoniae\) T1 - Tiefenlernen als Werkzeug zur Identifizierung und Strukturbestimmung neuer Proteinkomplexe aus der \(in\)-\(situ\)-Elektronenkryotomographie von \(Mycoplasma\) \(pneumoniae\) N2 - The holy grail of structural biology is to study a protein in situ, and this goal has been fast approaching since the resolution revolution and the achievement of atomic resolution. A cell's interior is not a dilute environment, and proteins have evolved to fold and function as needed in that environment; as such, an investigation of a cellular component should ideally include the full complexity of the cellular environment. Imaging whole cells in three dimensions using electron cryotomography is the best method to accomplish this goal, but it comes with a limitation on sample thickness and produces noisy data unamenable to direct analysis. This thesis establishes a novel workflow to systematically analyse whole-cell electron cryotomography data in three dimensions and to find and identify instances of protein complexes in the data to set up a determination of their structure and identity for success. Mycoplasma pneumoniae is a very small parasitic bacterium with fewer than 700 protein-coding genes, is thin enough and small enough to be imaged in large quantities by electron cryotomography, and can grow directly on the grids used for imaging, making it ideal for exploratory studies in structural proteomics. As part of the workflow, a methodology for training deep-learning-based particle-picking models is established. As a proof of principle, a dataset of whole-cell Mycoplasma pneumoniae tomograms is used with this workflow to characterize a novel membrane-associated complex observed in the data. Ultimately, 25431 such particles are picked from 353 tomograms and refined to a density map with a resolution of 11 Å. Making good use of orthogonal datasets to filter search space and verify results, structures were predicted for candidate proteins and checked for suitable fit in the density map. In the end, with this approach, nine proteins were found to be part of the complex, which appears to be associated with chaperone activity and interact with translocon machinery. Visual proteomics refers to the ultimate potential of in situ electron cryotomography: the comprehensive interpretation of tomograms. The workflow presented here is demonstrated to help in reaching that potential. N2 - Der heilige Gral der Strukturbiologie ist die Untersuchung eines Proteins in situ, und dieses Ziel ist seit der Auflösungsrevolution und dem Erreichen der atomaren Auflösung in greifbare Nähe gerückt. Das Innere einer Zelle ist keine verdünnte Umgebung, und Proteine haben sich so entwickelt, dass sie sich falten und so funktionieren, wie es in dieser Umgebung erforderlich ist; daher sollte die Untersuchung einer zellulären Komponente idealerweise die gesamte Komplexität der zellulären Umgebung umfassen. Die Abbildung ganzer Zellen in drei Dimensionen mit Hilfe der Elektronenkryotomographie ist die beste Methode, um dieses Ziel zu erreichen, aber sie ist mit einer Beschränkung der Probendicke verbunden und erzeugt verrauschte Daten, die sich nicht für eine direkte Analyse eignen. In dieser Dissertation wird ein neuartiger Workflow zur systematischen dreidimensionalen Analyse von Ganzzell-Elektronenkryotomographiedaten und zur Auffindung und Identifizierung von Proteinkomplexen in diesen Daten entwickelt, um eine erfolgreiche Bestimmung ihrer Struktur und Identität zu ermöglichen. Mycoplasma pneumoniae ist ein sehr kleines parasitäres Bakterium mit weniger als 700 proteinkodierenden Genen. Es ist dünn und klein genug, um in grossen Mengen durch Elektronenkryotomographie abgebildet zu werden, und kann direkt auf den für die Abbildung verwendeten Gittern wachsen, was es ideal für Sondierungsstudien in der strukturellen Proteomik macht. Als Teil des Workflows wird eine Methodik für das Training von Deep-Learning-basierten Partikelpicken-Modellen entwickelt. Als Proof-of-Principle wird ein Dataset von Ganzzell-Tomogrammen von Mycoplasma pneumoniae mit diesem Workflow verwendet, um einen neuartigen membranassoziierten Komplex zu charakterisieren, der in den Daten beobachtet wurde. Insgesamt wurden 25431 solcher Partikel aus 353 Tomogrammen gepickt und zu einer Dichtekarte mit einer Auflösung von 11 Å verfeinert. Unter Verwendung orthogonaler Datensätze zur Filterung des Suchraums und zur Überprüfung der Ergebnisse wurden Strukturen für Protein-Kandidaten vorhergesagt und auf ihre Eignung für die Dichtekarte überprüft. Letztendlich wurden mit diesem Ansatz neun Proteine als Bestandteile des Komplexes gefunden, der offenbar mit der Chaperonaktivität in Verbindung steht und mit der Translocon-Maschinerie interagiert. Das ultimative Potenzial der In-situ-Elektronenkryotomographie – die umfassende Interpretation von Tomogrammen – wird als visuelle Proteomik bezeichnet. Der hier vorgestellte Workflow soll dabei helfen, dieses Potenzial auszuschöpfen. KW - Kryoelektronenmikroskopie KW - Tomografie KW - Mycoplasma pneumoniae KW - Deep learning KW - cryo-EM KW - cryo-ET KW - tomography KW - mycoplasma KW - pneumoniae KW - deep learning KW - particle picking KW - membrane protein KW - visual proteomics Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-313447 ER - TY - RPRT A1 - Elsayed, Karim A1 - Rizk, Amr T1 - Response Times in Time-to-Live Caching Hierarchies under Random Network Delays T2 - Würzburg Workshop on Next-Generation Communication Networks (WueWoWas'22) N2 - Time-to-Live (TTL) caches decouple the occupancy of objects in cache through object-specific validity timers. Stateof- the art techniques provide exact methods for the calculation of object-specific hit probabilities given entire cache hierarchies with random inter-cache network delays. The system hit probability is a provider-centric metric as it relates to the origin offload, i.e., the decrease in the number of requests that are served by the content origin server. In this paper we consider a user-centric metric, i.e., the response time, which is shown to be structurally different from the system hit probability. Equipped with the state-of-theart exact modeling technique using Markov-arrival processes we derive expressions for the expected object response time and pave a way for its optimization under network delays. KW - Datennetz KW - TTL Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-280843 ER - TY - RPRT A1 - Alfredsson, Rebecka A1 - Kassler, Andreas A1 - Vestin, Jonathan A1 - Pieska, Marcus A1 - Amend, Markus T1 - Accelerating a Transport Layer based 5G Multi-Access Proxy on SmartNIC T2 - Würzburg Workshop on Next-Generation Communication Networks (WueWoWas'22) N2 - Utilizing multiple access technologies such as 5G, 4G, and Wi-Fi within a coherent framework is currently standardized by 3GPP within 5G ATSSS. Indeed, distributing packets over multiple networks can lead to increased robustness, resiliency and capacity. A key part of such a framework is the multi-access proxy, which transparently distributes packets over multiple paths. As the proxy needs to serve thousands of customers, scalability and performance are crucial for operator deployments. In this paper, we leverage recent advancements in data plane programming, implement a multi-access proxy based on the MP-DCCP tunneling approach in P4 and hardware accelerate it by deploying the pipeline on a smartNIC. This is challenging due to the complex scheduling and congestion control operations involved. We present our pipeline and data structures design for congestion control and packet scheduling state management. Initial measurements in our testbed show that packet latency is in the range of 25 μs demonstrating the feasibility of our approach. KW - Datennetz KW - multipath KW - MP-DCCP KW - 5G-ATSSS KW - networking KW - dataplane programming KW - P4 Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-280798 ER - TY - JOUR A1 - Bencurova, Elena A1 - Shityakov, Sergey A1 - Schaack, Dominik A1 - Kaltdorf, Martin A1 - Sarukhanyan, Edita A1 - Hilgarth, Alexander A1 - Rath, Christin A1 - Montenegro, Sergio A1 - Roth, Günter A1 - Lopez, Daniel A1 - Dandekar, Thomas T1 - Nanocellulose composites as smart devices with chassis, light-directed DNA Storage, engineered electronic properties, and chip integration JF - Frontiers in Bioengineering and Biotechnology N2 - The rapid development of green and sustainable materials opens up new possibilities in the field of applied research. Such materials include nanocellulose composites that can integrate many components into composites and provide a good chassis for smart devices. In our study, we evaluate four approaches for turning a nanocellulose composite into an information storage or processing device: 1) nanocellulose can be a suitable carrier material and protect information stored in DNA. 2) Nucleotide-processing enzymes (polymerase and exonuclease) can be controlled by light after fusing them with light-gating domains; nucleotide substrate specificity can be changed by mutation or pH change (read-in and read-out of the information). 3) Semiconductors and electronic capabilities can be achieved: we show that nanocellulose is rendered electronic by iodine treatment replacing silicon including microstructures. Nanocellulose semiconductor properties are measured, and the resulting potential including single-electron transistors (SET) and their properties are modeled. Electric current can also be transported by DNA through G-quadruplex DNA molecules; these as well as classical silicon semiconductors can easily be integrated into the nanocellulose composite. 4) To elaborate upon miniaturization and integration for a smart nanocellulose chip device, we demonstrate pH-sensitive dyes in nanocellulose, nanopore creation, and kinase micropatterning on bacterial membranes as well as digital PCR micro-wells. Future application potential includes nano-3D printing and fast molecular processors (e.g., SETs) integrated with DNA storage and conventional electronics. This would also lead to environment-friendly nanocellulose chips for information processing as well as smart nanocellulose composites for biomedical applications and nano-factories. KW - nanocellulose KW - DNA storage KW - light-gated proteins KW - single-electron transistors KW - protein chip Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-283033 SN - 2296-4185 VL - 10 ER - TY - JOUR A1 - Krenzer, Adrian A1 - Makowski, Kevin A1 - Hekalo, Amar A1 - Fitting, Daniel A1 - Troya, Joel A1 - Zoller, Wolfram G. A1 - Hann, Alexander A1 - Puppe, Frank T1 - Fast machine learning annotation in the medical domain: a semi-automated video annotation tool for gastroenterologists JF - BioMedical Engineering OnLine N2 - Background Machine learning, especially deep learning, is becoming more and more relevant in research and development in the medical domain. For all the supervised deep learning applications, data is the most critical factor in securing successful implementation and sustaining the progress of the machine learning model. Especially gastroenterological data, which often involves endoscopic videos, are cumbersome to annotate. Domain experts are needed to interpret and annotate the videos. To support those domain experts, we generated a framework. With this framework, instead of annotating every frame in the video sequence, experts are just performing key annotations at the beginning and the end of sequences with pathologies, e.g., visible polyps. Subsequently, non-expert annotators supported by machine learning add the missing annotations for the frames in-between. Methods In our framework, an expert reviews the video and annotates a few video frames to verify the object’s annotations for the non-expert. In a second step, a non-expert has visual confirmation of the given object and can annotate all following and preceding frames with AI assistance. After the expert has finished, relevant frames will be selected and passed on to an AI model. This information allows the AI model to detect and mark the desired object on all following and preceding frames with an annotation. Therefore, the non-expert can adjust and modify the AI predictions and export the results, which can then be used to train the AI model. Results Using this framework, we were able to reduce workload of domain experts on average by a factor of 20 on our data. This is primarily due to the structure of the framework, which is designed to minimize the workload of the domain expert. Pairing this framework with a state-of-the-art semi-automated AI model enhances the annotation speed further. Through a prospective study with 10 participants, we show that semi-automated annotation using our tool doubles the annotation speed of non-expert annotators compared to a well-known state-of-the-art annotation tool. Conclusion In summary, we introduce a framework for fast expert annotation for gastroenterologists, which reduces the workload of the domain expert considerably while maintaining a very high annotation quality. The framework incorporates a semi-automated annotation system utilizing trained object detection models. The software and framework are open-source. KW - object detection KW - machine learning KW - deep learning KW - annotation KW - endoscopy KW - gastroenterology KW - automation Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-300231 VL - 21 IS - 1 ER - TY - JOUR A1 - Kaltdorf, Kristin Verena A1 - Schulze, Katja A1 - Helmprobst, Frederik A1 - Kollmannsberger, Philip A1 - Dandekar, Thomas A1 - Stigloher, Christian T1 - Fiji macro 3D ART VeSElecT: 3D automated reconstruction tool for vesicle structures of electron tomograms JF - PLoS Computational Biology N2 - Automatic image reconstruction is critical to cope with steadily increasing data from advanced microscopy. We describe here the Fiji macro 3D ART VeSElecT which we developed to study synaptic vesicles in electron tomograms. We apply this tool to quantify vesicle properties (i) in embryonic Danio rerio 4 and 8 days past fertilization (dpf) and (ii) to compare Caenorhabditis elegans N2 neuromuscular junctions (NMJ) wild-type and its septin mutant (unc-59(e261)). We demonstrate development-specific and mutant-specific changes in synaptic vesicle pools in both models. We confirm the functionality of our macro by applying our 3D ART VeSElecT on zebrafish NMJ showing smaller vesicles in 8 dpf embryos then 4 dpf, which was validated by manual reconstruction of the vesicle pool. Furthermore, we analyze the impact of C. elegans septin mutant unc-59(e261) on vesicle pool formation and vesicle size. Automated vesicle registration and characterization was implemented in Fiji as two macros (registration and measurement). This flexible arrangement allows in particular reducing false positives by an optional manual revision step. Preprocessing and contrast enhancement work on image-stacks of 1nm/pixel in x and y direction. Semi-automated cell selection was integrated. 3D ART VeSElecT removes interfering components, detects vesicles by 3D segmentation and calculates vesicle volume and diameter (spherical approximation, inner/outer diameter). Results are collected in color using the RoiManager plugin including the possibility of manual removal of non-matching confounder vesicles. Detailed evaluation considered performance (detected vesicles) and specificity (true vesicles) as well as precision and recall. We furthermore show gain in segmentation and morphological filtering compared to learning based methods and a large time gain compared to manual segmentation. 3D ART VeSElecT shows small error rates and its speed gain can be up to 68 times faster in comparison to manual annotation. Both automatic and semi-automatic modes are explained including a tutorial. KW - Biology KW - Vesicles KW - Caenorhabditis elegans KW - Zebrafish KW - Septins KW - Synaptic vesicles KW - Neuromuscular junctions KW - Computer software KW - Synapses Y1 - 2017 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-172112 VL - 13 IS - 1 ER - TY - JOUR A1 - Döllinger, Nina A1 - Wienrich, Carolin A1 - Latoschik, Marc Erich T1 - Challenges and opportunities of immersive technologies for mindfulness meditation: a systematic review JF - Frontiers in Virtual Reality N2 - Mindfulness is considered an important factor of an individual's subjective well-being. Consequently, Human-Computer Interaction (HCI) has investigated approaches that strengthen mindfulness, i.e., by inventing multimedia technologies to support mindfulness meditation. These approaches often use smartphones, tablets, or consumer-grade desktop systems to allow everyday usage in users' private lives or in the scope of organized therapies. Virtual, Augmented, and Mixed Reality (VR, AR, MR; in short: XR) significantly extend the design space for such approaches. XR covers a wide range of potential sensory stimulation, perceptive and cognitive manipulations, content presentation, interaction, and agency. These facilities are linked to typical XR-specific perceptions that are conceptually closely related to mindfulness research, such as (virtual) presence and (virtual) embodiment. However, a successful exploitation of XR that strengthens mindfulness requires a systematic analysis of the potential interrelation and influencing mechanisms between XR technology, its properties, factors, and phenomena and existing models and theories of the construct of mindfulness. This article reports such a systematic analysis of XR-related research from HCI and life sciences to determine the extent to which existing research frameworks on HCI and mindfulness can be applied to XR technologies, the potential of XR technologies to support mindfulness, and open research gaps. Fifty papers of ACM Digital Library and National Institutes of Health's National Library of Medicine (PubMed) with and without empirical efficacy evaluation were included in our analysis. The results reveal that at the current time, empirical research on XR-based mindfulness support mainly focuses on therapy and therapeutic outcomes. Furthermore, most of the currently investigated XR-supported mindfulness interactions are limited to vocally guided meditations within nature-inspired virtual environments. While an analysis of empirical research on those systems did not reveal differences in mindfulness compared to non-mediated mindfulness practices, various design proposals illustrate that XR has the potential to provide interactive and body-based innovations for mindfulness practice. We propose a structured approach for future work to specify and further explore the potential of XR as mindfulness-support. The resulting framework provides design guidelines for XR-based mindfulness support based on the elements and psychological mechanisms of XR interactions. KW - virtual reality KW - augmented reality KW - mindfulness KW - XR KW - meditation Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-259047 VL - 2 ER - TY - JOUR A1 - Prakash, Subash A1 - Unnikrishnan, Vishnu A1 - Pryss, Rüdiger A1 - Kraft, Robin A1 - Schobel, Johannes A1 - Hannemann, Ronny A1 - Langguth, Berthold A1 - Schlee, Winfried A1 - Spiliopoulou, Myra T1 - Interactive system for similarity-based inspection and assessment of the well-being of mHealth users JF - Entropy N2 - Recent digitization technologies empower mHealth users to conveniently record their Ecological Momentary Assessments (EMA) through web applications, smartphones, and wearable devices. These recordings can help clinicians understand how the users' condition changes, but appropriate learning and visualization mechanisms are required for this purpose. We propose a web-based visual analytics tool, which processes clinical data as well as EMAs that were recorded through a mHealth application. The goals we pursue are (1) to predict the condition of the user in the near and the far future, while also identifying the clinical data that mostly contribute to EMA predictions, (2) to identify users with outlier EMA, and (3) to show to what extent the EMAs of a user are in line with or diverge from those users similar to him/her. We report our findings based on a pilot study on patient empowerment, involving tinnitus patients who recorded EMAs with the mHealth app TinnitusTips. To validate our method, we also derived synthetic data from the same pilot study. Based on this setting, results for different use cases are reported. KW - medical analytics KW - condition prediction KW - ecological momentary assessment KW - visual analytics KW - time series Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-252333 SN - 1099-4300 VL - 23 IS - 12 ER - TY - JOUR A1 - Kraft, Robin A1 - Birk, Ferdinand A1 - Reichert, Manfred A1 - Deshpande, Aniruddha A1 - Schlee, Winfried A1 - Langguth, Berthold A1 - Baumeister, Harald A1 - Probst, Thomas A1 - Spiliopoulou, Myra A1 - Pryss, Rüdiger T1 - Efficient processing of geospatial mHealth data using a scalable crowdsensing platform JF - Sensors N2 - Smart sensors and smartphones are becoming increasingly prevalent. Both can be used to gather environmental data (e.g., noise). Importantly, these devices can be connected to each other as well as to the Internet to collect large amounts of sensor data, which leads to many new opportunities. In particular, mobile crowdsensing techniques can be used to capture phenomena of common interest. Especially valuable insights can be gained if the collected data are additionally related to the time and place of the measurements. However, many technical solutions still use monolithic backends that are not capable of processing crowdsensing data in a flexible, efficient, and scalable manner. In this work, an architectural design was conceived with the goal to manage geospatial data in challenging crowdsensing healthcare scenarios. It will be shown how the proposed approach can be used to provide users with an interactive map of environmental noise, allowing tinnitus patients and other health-conscious people to avoid locations with harmful sound levels. Technically, the shown approach combines cloud-native applications with Big Data and stream processing concepts. In general, the presented architectural design shall serve as a foundation to implement practical and scalable crowdsensing platforms for various healthcare scenarios beyond the addressed use case. KW - mHealth KW - crowdsensing KW - tinnitus KW - geospatial data KW - cloud-native KW - stream processing KW - scalability KW - architectural design Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-207826 SN - 1424-8220 VL - 20 IS - 12 ER - TY - JOUR A1 - Klemz, Boris A1 - Rote, Günter T1 - Linear-Time Algorithms for Maximum-Weight Induced Matchings and Minimum Chain Covers in Convex Bipartite Graphs JF - Algorithmica N2 - A bipartite graph G=(U,V,E) is convex if the vertices in V can be linearly ordered such that for each vertex u∈U, the neighbors of u are consecutive in the ordering of V. An induced matching H of G is a matching for which no edge of E connects endpoints of two different edges of H. We show that in a convex bipartite graph with n vertices and m weighted edges, an induced matching of maximum total weight can be computed in O(n+m) time. An unweighted convex bipartite graph has a representation of size O(n) that records for each vertex u∈U the first and last neighbor in the ordering of V. Given such a compact representation, we compute an induced matching of maximum cardinality in O(n) time. In convex bipartite graphs, maximum-cardinality induced matchings are dual to minimum chain covers. A chain cover is a covering of the edge set by chain subgraphs, that is, subgraphs that do not contain induced matchings of more than one edge. Given a compact representation, we compute a representation of a minimum chain cover in O(n) time. If no compact representation is given, the cover can be computed in O(n+m) time. All of our algorithms achieve optimal linear running time for the respective problem and model, and they improve and generalize the previous results in several ways: The best algorithms for the unweighted problem versions had a running time of O(n\(^{2}\)) (Brandstädt et al. in Theor. Comput. Sci. 381(1–3):260–265, 2007. https://doi.org/10.1016/j.tcs.2007.04.006). The weighted case has not been considered before. KW - dynamic programming KW - graph algorithm KW - induced matching KW - chain cover KW - convex bipartite graph KW - certifying algorithm Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-267876 SN - 1432-0541 VL - 84 IS - 4 ER - TY - JOUR A1 - Davidson, Padraig A1 - Düking, Peter A1 - Zinner, Christoph A1 - Sperlich, Billy A1 - Hotho, Andreas T1 - Smartwatch-Derived Data and Machine Learning Algorithms Estimate Classes of Ratings of Perceived Exertion in Runners: A Pilot Study JF - Sensors N2 - The rating of perceived exertion (RPE) is a subjective load marker and may assist in individualizing training prescription, particularly by adjusting running intensity. Unfortunately, RPE has shortcomings (e.g., underreporting) and cannot be monitored continuously and automatically throughout a training sessions. In this pilot study, we aimed to predict two classes of RPE (≤15 “Somewhat hard to hard” on Borg’s 6–20 scale vs. RPE >15 in runners by analyzing data recorded by a commercially-available smartwatch with machine learning algorithms. Twelve trained and untrained runners performed long-continuous runs at a constant self-selected pace to volitional exhaustion. Untrained runners reported their RPE each kilometer, whereas trained runners reported every five kilometers. The kinetics of heart rate, step cadence, and running velocity were recorded continuously ( 1 Hz ) with a commercially-available smartwatch (Polar V800). We trained different machine learning algorithms to estimate the two classes of RPE based on the time series sensor data derived from the smartwatch. Predictions were analyzed in different settings: accuracy overall and per runner type; i.e., accuracy for trained and untrained runners independently. We achieved top accuracies of 84.8 % for the whole dataset, 81.8 % for the trained runners, and 86.1 % for the untrained runners. We predict two classes of RPE with high accuracy using machine learning and smartwatch data. This approach might aid in individualizing training prescriptions. KW - artificial intelligence KW - endurance KW - exercise intensity KW - precision training KW - prediction KW - wearable Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-205686 SN - 1424-8220 VL - 20 IS - 9 ER - TY - RPRT ED - Hoßfeld, Tobias ED - Wunderer, Stefan T1 - White Paper on Crowdsourced Network and QoE Measurements – Definitions, Use Cases and Challenges N2 - The goal of the white paper at hand is as follows. The definitions of the terms build a framework for discussions around the hype topic ‘crowdsourcing’. This serves as a basis for differentiation and a consistent view from different perspectives on crowdsourced network measurements, with the goal to provide a commonly accepted definition in the community. The focus is on the context of mobile and fixed network operators, but also on measurements of different layers (network, application, user layer). In addition, the white paper shows the value of crowdsourcing for selected use cases, e.g., to improve QoE or regulatory issues. Finally, the major challenges and issues for researchers and practitioners are highlighted. This white paper is the outcome of the Würzburg seminar on “Crowdsourced Network and QoE Measurements” which took place from 25-26 September 2019 in Würzburg, Germany. International experts were invited from industry and academia. They are well known in their communities, having different backgrounds in crowdsourcing, mobile networks, network measurements, network performance, Quality of Service (QoS), and Quality of Experience (QoE). The discussions in the seminar focused on how crowdsourcing will support vendors, operators, and regulators to determine the Quality of Experience in new 5G networks that enable various new applications and network architectures. As a result of the discussions, the need for a white paper manifested, with the goal of providing a scientific discussion of the terms “crowdsourced network measurements” and “crowdsourced QoE measurements”, describing relevant use cases for such crowdsourced data, and its underlying challenges. During the seminar, those main topics were identified, intensively discussed in break-out groups, and brought back into the plenum several times. The outcome of the seminar is this white paper at hand which is – to our knowledge – the first one covering the topic of crowdsourced network and QoE measurements. KW - Crowdsourcing KW - Network Measurements KW - Quality of Service (QoS) KW - Quality of Experience (QoE) KW - crowdsourced network measurements KW - crowdsourced QoE measurements Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-202327 ER - TY - JOUR A1 - Pawellek, Ruben A1 - Krmar, Jovana A1 - Leistner, Adrian A1 - Djajić, Nevena A1 - Otašević, Biljana A1 - Protić, Ana A1 - Holzgrabe, Ulrike T1 - Charged aerosol detector response modeling for fatty acids based on experimental settings and molecular features: a machine learning approach JF - Journal of Cheminformatics N2 - The charged aerosol detector (CAD) is the latest representative of aerosol-based detectors that generate a response independent of the analytes' chemical structure. This study was aimed at accurately predicting the CAD response of homologous fatty acids under varying experimental conditions. Fatty acids from C12 to C18 were used as model substances due to semivolatile characterics that caused non-uniform CAD behaviour. Considering both experimental conditions and molecular descriptors, a mixed quantitative structure-property relationship (QSPR) modeling was performed using Gradient Boosted Trees (GBT). The ensemble of 10 decisions trees (learning rate set at 0.55, the maximal depth set at 5, and the sample rate set at 1.0) was able to explain approximately 99% (Q\(^2\): 0.987, RMSE: 0.051) of the observed variance in CAD responses. Validation using an external test compound confirmed the high predictive ability of the model established (R-2: 0.990, RMSEP: 0.050). With respect to the intrinsic attribute selection strategy, GBT used almost all independent variables during model building. Finally, it attributed the highest importance to the power function value, the flow rate of the mobile phase, evaporation temperature, the content of the organic solvent in the mobile phase and the molecular descriptors such as molecular weight (MW), Radial Distribution Function-080/weighted by mass (RDF080m) and average coefficient of the last eigenvector from distance/detour matrix (Ve2_D/Dt). The identification of the factors most relevant to the CAD responsiveness has contributed to a better understanding of the underlying mechanisms of signal generation. An increased CAD response that was obtained for acetone as organic modifier demonstrated its potential to replace the more expensive and environmentally harmful acetonitrile. KW - High-performance liquid chromatography (HPLC) KW - Charged aerosol detector (CAD) KW - Gradient boosted trees (GBT) KW - Quantitative structure-property relationship modeling (QSPR) KW - Fatty acids Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-261618 VL - 13 IS - 1 ER - TY - JOUR A1 - Unruh, Fabian A1 - Landeck, Maximilian A1 - Oberdörfer, Sebastian A1 - Lugrin, Jean-Luc A1 - Latoschik, Marc Erich T1 - The Influence of Avatar Embodiment on Time Perception - Towards VR for Time-Based Therapy JF - Frontiers in Virtual Reality N2 - Psycho-pathological conditions, such as depression or schizophrenia, are often accompanied by a distorted perception of time. People suffering from this conditions often report that the passage of time slows down considerably and that they are “stuck in time.” Virtual Reality (VR) could potentially help to diagnose and maybe treat such mental conditions. However, the conditions in which a VR simulation could correctly diagnose a time perception deviation are still unknown. In this paper, we present an experiment investigating the difference in time experience with and without a virtual body in VR, also known as avatar. The process of substituting a person’s body with a virtual body is called avatar embodiment. Numerous studies demonstrated interesting perceptual, emotional, behavioral, and psychological effects caused by avatar embodiment. However, the relations between time perception and avatar embodiment are still unclear. Whether or not the presence or absence of an avatar is already influencing time perception is still open to question. Therefore, we conducted a between-subjects design with and without avatar embodiment as well as a real condition (avatar vs. no-avatar vs. real). A group of 105 healthy subjects had to wait for seven and a half minutes in a room without any distractors (e.g., no window, magazine, people, decoration) or time indicators (e.g., clocks, sunlight). The virtual environment replicates the real physical environment. Participants were unaware that they will be asked to estimate their waiting time duration as well as describing their experience of the passage of time at a later stage. Our main finding shows that the presence of an avatar is leading to a significantly faster perceived passage of time. It seems to be promising to integrate avatar embodiment in future VR time-based therapy applications as they potentially could modulate a user’s perception of the passage of time. We also found no significant difference in time perception between the real and the VR conditions (avatar, no-avatar), but further research is needed to better understand this outcome. KW - virtual reality KW - time perception KW - avatar embodiment KW - immersion KW - human computer interaction (HCI) Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-259076 VL - 2 ER - TY - THES A1 - Huber, Stephan T1 - Proxemo: Documenting Observed Emotions in HCI T1 - Proxemo: Die Dokumentation Beobachteter Emotionen in der Mensch-Computer-Interaktion N2 - For formative evaluations of user experience (UX) a variety of methods have been developed over the years. However, most techniques require the users to interact with the study as a secondary task. This active involvement in the evaluation is not inclusive of all users and potentially biases the experience currently being studied. Yet there is a lack of methods for situations in which the user has no spare cognitive resources. This condition occurs when 1) users' cognitive abilities are impaired (e.g., people with dementia) or 2) users are confronted with very demanding tasks (e.g., air traffic controllers). In this work we focus on emotions as a key component of UX and propose the new structured observation method Proxemo for formative UX evaluations. Proxemo allows qualified observers to document users' emotions by proxy in real time and then directly link them to triggers. Technically this is achieved by synchronising the timestamps of emotions documented by observers with a video recording of the interaction. In order to facilitate the documentation of observed emotions in highly diverse contexts we conceptualise and implement two separate versions of a documentation aid named Proxemo App. For formative UX evaluations of technology-supported reminiscence sessions with people with dementia, we create a smartwatch app to discreetly document emotions from the categories anger, general alertness, pleasure, wistfulness and pride. For formative UX evaluations of prototypical user interfaces with air traffic controllers we create a smartphone app to efficiently document emotions from the categories anger, boredom, surprise, stress and pride. Descriptive case studies in both application domains indicate the feasibility and utility of the method Proxemo and the appropriateness of the respectively adapted design of the Proxemo App. The third part of this work is a series of meta-evaluation studies to determine quality criteria of Proxemo. We evaluate Proxemo regarding its reliability, validity, thoroughness and effectiveness, and compare Proxemo's efficiency and the observers' experience to documentation with pen and paper. Proxemo is reliable, as well as more efficient, thorough and effective than handwritten notes and provides a better UX to observers. Proxemo compares well with existing methods where benchmarks are available. With Proxemo we contribute a validated structured observation method that has shown to meet requirements formative UX evaluations in the extreme contexts of users with cognitive impairments or high task demands. Proxemo is agnostic regarding researchers' theoretical approaches and unites reductionist and holistic perspectives within one method. Future work should explore the applicability of Proxemo for further domains and extend the list of audited quality criteria to include, for instance, downstream utility. With respect to basic research we strive to better understand the sources leading observers to empathic judgments and propose reminisce and older adults as model environment for investigating mixed emotions. N2 - Für formative Evaluationen der User Experience (UX) wurden im Laufe der Jahre zahlreiche Methoden entwickelt. Die meisten Methoden erfordern jedoch, dass die Benutzer als Nebenaufgabe mit der Studie interagieren. Diese aktive Beteiligung an der Evaluation kann das untersuchte Erlebnis verfälschen und schließt Benutzer komplett aus, die keine kognitiven Ressourcen zur Verfügung haben. Dies ist der Fall, wenn 1) die kognitiven Fähigkeiten der Benutzer beeinträchtigt sind (z. B. Menschen mit Demenz) oder 2) Benutzer mit sehr anspruchsvollen Aufgaben konfrontiert sind (z. B. Fluglotsen). In dieser Arbeit konzentrieren wir uns auf Emotionen als eine Schlüsselkomponente von UX und schlagen die neue strukturierte Beobachtungsmethode Proxemo für formative UX-Evaluationen vor. Proxemo ermöglicht es qualifizierten Beobachtern, die Emotionen der Nutzer in Echtzeit zu dokumentieren und sie direkt mit Auslösern zu verknüpfen. Technisch wird dies erreicht, indem die Zeitstempel der von den Beobachtern dokumentierten Emotionen mit einer Videoaufzeichnung der Interaktion synchronisiert werden. Um die Dokumentation von beobachteten Emotionen in sehr unterschiedlichen Kontexten zu erleichtern, konzipieren und implementieren wir zwei verschiedene Versionen einer Dokumentationshilfe namens Proxemo App. Für formative UX-Evaluationen von technologiegestützten Erinnerungssitzungen mit Menschen mit Demenz erstellen wir eine Smartwatch-App zur unauffälligen Dokumentation von Emotionen aus den Kategorien Ärger, allgemeine Wachsamkeit, Freude, Wehmut und Stolz. Für formative UX-Evaluationen prototypischer Nutzerschnittstellen mit Fluglotsen erstellen wir eine Smartphone-App zur effizienten Dokumentation von Emotionen aus den Kategorien Ärger, Langeweile, Überraschung, Stress und Stolz. Deskriptive Fallstudien in beiden Anwendungsfeldern zeigen die Machbarkeit und den Nutzen der Methode Proxemo und die Angemessenheit des jeweiligen Designs der Proxemo App. Der dritte Teil dieser Arbeit besteht aus einer Reihe von Meta-Evaluationsstudien zu den Gütekriterien von Proxemo. Wir evaluieren Proxemo hinsichtlich der Reliabilität, Validität, Gründlichkeit und Effektivität, und vergleichen die Effizienz von Proxemo und die UX der Beobachter mit der Dokumentation mit Stift und Papier. Proxemo ist reliabel, sowie effizienter, gründlicher und effektiver als handschriftliche Notizen und bietet den Beobachtern eine bessere UX. Proxemo schneidet gut ab im Vergleich zu bestehenden Methoden, für die Benchmarks verfügbar sind. Mit Proxemo stellen wir eine validierte, strukturierte Beobachtungsmethode vor, die nachweislich den Anforderungen formativer UX Evaluationen in den extremen Kontexten von Benutzern mit kognitiven Beeinträchtigungen oder hohen Aufgabenanforderungen gerecht wird. Proxemo ist agnostisch bezüglich der theoretischen Ansätze von Forschenden und vereint reduktionistische und ganzheitliche Perspektiven in einer Methode. Zukünftige Arbeiten sollten die Anwendbarkeit von Proxemo für weitere Domänen erkunden und die Liste der geprüften Gütekriterien erweitern, zum Beispiel um das Kriterium Downstream Utility. In Bezug auf die Grundlagenforschung werden wir versuchen, die Quellen besser zu verstehen, auf denen die empathischen Urteile der Beobachter fußen und schlagen Erinnerungen und ältere Erwachsene als Modellumgebung für die künftige Erforschung gemischter Emotionen vor. KW - Gefühl KW - Wissenschaftliche Beobachtung KW - Methode KW - Benutzererlebnis KW - Benutzerforschung KW - Emotionserkennung KW - Emotion inference KW - Emotionsinterpretation Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-305730 ER - TY - THES A1 - Geißler, Stefan T1 - Performance Evaluation of Next-Generation Data Plane Architectures and their Components T1 - Leistungsbewertung von Data Plane Architekturen der Nächsten Generation sowie ihrer Einzelkomponenten N2 - In this doctoral thesis we cover the performance evaluation of next generation data plane architectures, comprised of complex software as well as programmable hardware components that allow fine granular configuration. In the scope of the thesis we propose mechanisms to monitor the performance of singular components and model key performance indicators of software based packet processing solutions. We present novel approaches towards network abstraction that allow the integration of heterogeneous data plane technologies into a singular network while maintaining total transparency between control and data plane. Finally, we investigate a full, complex system consisting of multiple software-based solutions and perform a detailed performance analysis. We employ simulative approaches to investigate overload control mechanisms that allow efficient operation under adversary conditions. The contributions of this work build the foundation for future research in the areas of network softwarization and network function virtualization. N2 - Diese Doktorarbeit behandelt die Leistungsbewertung von Data Plane Architekturen der nächsten Generation, die aus komplexen Softwarelösungen sowie programmierbaren Hardwarekomponenten bestehen. Hierbei werden Mechanismen entwickelt, die es ermöglichen, die Leistungsfähigkeit einzelner Komponenten zu messen und zentrale Leistungsindikatoren softwarebasierter Systeme zur Verarbeitung von Datenpaketen zu modellieren. Es werden neuartige Ansätze zur Netzabstraktion entworfen, die eine vollständig transparente Integration heterogener Technologien im selben Netz ermöglichen. Schließlich wird eine umfassende Leistungsbewertung eines komplexen Systems, das aus einer Vielzahl softwarebasierter Netzfunktionen besteht, durchgeführt. Anhand simulativer Modelle werden Überlastkontrollmechanismen entwickelt, die es dem System erlauben auch unter Überlast effizient zu arbeiten. Die Beiträge dieser Arbeit bilden die Grundlage weiterer Forschungen im Bereich der Softwarisierung von Netzen sowie der Virtualisierung von Netzfunktionen. T3 - Würzburger Beiträge zur Leistungsbewertung Verteilter Systeme - 02/21 KW - Leistungsbewertung KW - Simulation KW - Zeitdiskretes System KW - Implementierung KW - performance evaluation KW - simulation KW - discrete-time analysis KW - network softwarization KW - mobile networks Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-260157 SN - 1432-8801 ER - TY - JOUR A1 - Kammerer, Klaus A1 - Pryss, Rüdiger A1 - Hoppenstedt, Burkhard A1 - Sommer, Kevin A1 - Reichert, Manfred T1 - Process-driven and flow-based processing of industrial sensor data JF - Sensors N2 - For machine manufacturing companies, besides the production of high quality and reliable machines, requirements have emerged to maintain machine-related aspects through digital services. The development of such services in the field of the Industrial Internet of Things (IIoT) is dealing with solutions such as effective condition monitoring and predictive maintenance. However, appropriate data sources are needed on which digital services can be technically based. As many powerful and cheap sensors have been introduced over the last years, their integration into complex machines is promising for developing digital services for various scenarios. It is apparent that for components handling recorded data of these sensors they must usually deal with large amounts of data. In particular, the labeling of raw sensor data must be furthered by a technical solution. To deal with these data handling challenges in a generic way, a sensor processing pipeline (SPP) was developed, which provides effective methods to capture, process, store, and visualize raw sensor data based on a processing chain. Based on the example of a machine manufacturing company, the SPP approach is presented in this work. For the company involved, the approach has revealed promising results. KW - data stream processing KW - cyber-physical systems KW - processing pipeline KW - sensor networks Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-213089 SN - 1424-8220 VL - 20 IS - 18 ER - TY - JOUR A1 - Wick, Christoph A1 - Hartelt, Alexander A1 - Puppe, Frank T1 - Staff, symbol and melody detection of Medieval manuscripts written in square notation using deep Fully Convolutional Networks JF - Applied Sciences N2 - Even today, the automatic digitisation of scanned documents in general, but especially the automatic optical music recognition (OMR) of historical manuscripts, still remains an enormous challenge, since both handwritten musical symbols and text have to be identified. This paper focuses on the Medieval so-called square notation developed in the 11th–12th century, which is already composed of staff lines, staves, clefs, accidentals, and neumes that are roughly spoken connected single notes. The aim is to develop an algorithm that captures both the neumes, and in particular its melody, which can be used to reconstruct the original writing. Our pipeline is similar to the standard OMR approach and comprises a novel staff line and symbol detection algorithm based on deep Fully Convolutional Networks (FCN), which perform pixel-based predictions for either staff lines or symbols and their respective types. Then, the staff line detection combines the extracted lines to staves and yields an F\(_1\) -score of over 99% for both detecting lines and complete staves. For the music symbol detection, we choose a novel approach that skips the step to identify neumes and instead directly predicts note components (NCs) and their respective affiliation to a neume. Furthermore, the algorithm detects clefs and accidentals. Our algorithm predicts the symbol sequence of a staff with a diplomatic symbol accuracy rate (dSAR) of about 87%, which includes symbol type and location. If only the NCs without their respective connection to a neume, all clefs and accidentals are of interest, the algorithm reaches an harmonic symbol accuracy rate (hSAR) of approximately 90%. In general, the algorithm recognises a symbol in the manuscript with an F\(_1\) -score of over 96%. KW - optical music recognition KW - historical document analysis KW - medieval manuscripts KW - neume notation KW - fully convolutional neural networks Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-197248 SN - 2076-3417 VL - 9 IS - 13 ER - TY - RPRT A1 - Metzger, Florian T1 - Crowdsensed QoE for the community - a concept to make QoE assessment accessible N2 - In recent years several community testbeds as well as participatory sensing platforms have successfully established themselves to provide open data to everyone interested. Each of them with a specific goal in mind, ranging from collecting radio coverage data up to environmental and radiation data. Such data can be used by the community in their decision making, whether to subscribe to a specific mobile phone service that provides good coverage in an area or in finding a sunny and warm region for the summer holidays. However, the existing platforms are usually limiting themselves to directly measurable network QoS. If such a crowdsourced data set provides more in-depth derived measures, this would enable an even better decision making. A community-driven crowdsensing platform that derives spatial application-layer user experience from resource-friendly bandwidth estimates would be such a case, video streaming services come to mind as a prime example. In this paper we present a concept for such a system based on an initial prototype that eases the collection of data necessary to determine mobile-specific QoE at large scale. In addition we reason why the simple quality metric proposed here can hold its own. KW - Quality of Experience KW - Crowdsourcing KW - Crowdsensing KW - QoE Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-203748 N1 - Originally written in 2017, but never published. ER - TY - JOUR A1 - Seufert, Anika A1 - Schröder, Svenja A1 - Seufert, Michael T1 - Delivering User Experience over Networks: Towards a Quality of Experience Centered Design Cycle for Improved Design of Networked Applications JF - SN Computer Science N2 - To deliver the best user experience (UX), the human-centered design cycle (HCDC) serves as a well-established guideline to application developers. However, it does not yet cover network-specific requirements, which become increasingly crucial, as most applications deliver experience over the Internet. The missing network-centric view is provided by Quality of Experience (QoE), which could team up with UX towards an improved overall experience. By considering QoE aspects during the development process, it can be achieved that applications become network-aware by design. In this paper, the Quality of Experience Centered Design Cycle (QoE-CDC) is proposed, which provides guidelines on how to design applications with respect to network-specific requirements and QoE. Its practical value is showcased for popular application types and validated by outlining the design of a new smartphone application. We show that combining HCDC and QoE-CDC will result in an application design, which reaches a high UX and avoids QoE degradation. KW - user experience KW - human-centered design KW - design cycle KW - application design KW - quality of experience Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-271762 SN - 2661-8907 VL - 2 IS - 6 ER - TY - JOUR A1 - Krupitzer, Christian A1 - Eberhardinger, Benedikt A1 - Gerostathopoulos, Ilias A1 - Raibulet, Claudia T1 - Introduction to the special issue “Applications in Self-Aware Computing Systems and their Evaluation” JF - Computers N2 - The joint 1st Workshop on Evaluations and Measurements in Self-Aware Computing Systems (EMSAC 2019) and Workshop on Self-Aware Computing (SeAC) was held as part of the FAS* conference alliance in conjunction with the 16th IEEE International Conference on Autonomic Computing (ICAC) and the 13th IEEE International Conference on Self-Adaptive and Self-Organizing Systems (SASO) in Umeå, Sweden on 20 June 2019. The goal of this one-day workshop was to bring together researchers and practitioners from academic environments and from the industry to share their solutions, ideas, visions, and doubts in self-aware computing systems in general and in the evaluation and measurements of such systems in particular. The workshop aimed to enable discussions, partnerships, and collaborations among the participants. This special issue follows the theme of the workshop. It contains extended versions of workshop presentations as well as additional contributions. KW - self-aware computing systems KW - quality evaluation KW - measurements KW - quality assurance KW - autonomous KW - self-adaptive KW - self-managing systems Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-203439 SN - 2073-431X VL - 9 IS - 1 ER - TY - THES A1 - Peng, Dongliang T1 - An Optimization-Based Approach for Continuous Map Generalization T1 - Optimierung für die kontinuierliche Generalisierung von Landkarten N2 - Maps are the main tool to represent geographical information. Geographical information is usually scale-dependent, so users need to have access to maps at different scales. In our digital age, the access is realized by zooming. As discrete changes during the zooming tend to distract users, smooth changes are preferred. This is why some digital maps are trying to make the zooming as continuous as they can. The process of producing maps at different scales with smooth changes is called continuous map generalization. In order to produce maps of high quality, cartographers often take into account additional requirements. These requirements are transferred to models in map generalization. Optimization for map generalization is important not only because it finds optimal solutions in the sense of the models, but also because it helps us to evaluate the quality of the models. Optimization, however, becomes more delicate when we deal with continuous map generalization. In this area, there are requirements not only for a specific map but also for relations between maps at difference scales. This thesis is about continuous map generalization based on optimization. First, we show the background of our research topics. Second, we find optimal sequences for aggregating land-cover areas. We compare the A$^{\!\star}$\xspace algorithm and integer linear programming in completing this task. Third, we continuously generalize county boundaries to provincial boundaries based on compatible triangulations. We morph between the two sets of boundaries, using dynamic programming to compute the correspondence. Fourth, we continuously generalize buildings to built-up areas by aggregating and growing. In this work, we group buildings with the help of a minimum spanning tree. Fifth, we define vertex trajectories that allow us to morph between polylines. We require that both the angles and the edge lengths change linearly over time. As it is impossible to fulfill all of these requirements simultaneously, we mediate between them using least-squares adjustment. Sixth, we discuss the performance of some commonly used data structures for a specific spatial problem. Seventh, we conclude this thesis and present open problems. N2 - Maps are the main tool to represent geographical information. Users often zoom in and out to access maps at different scales. Continuous map generalization tries to make the changes between different scales smooth, which is essential to provide users with comfortable zooming experience. In order to achieve continuous map generalization with high quality, we optimize some important aspects of maps. In this book, we have used optimization in the generalization of land-cover areas, administrative boundaries, buildings, and coastlines. According to our experiments, continuous map generalization indeed benefits from optimization. N2 - Landkarten sind das wichtigste Werkzeug zur Repräsentation geografischer Information. Unter der Generalisierung von Landkarten versteht man die Aufbereitung von geografischen Informationen aus detaillierten Daten zur Generierung von kleinmaßstäbigen Karten. Nutzer von Online-Karten zoomen oft in eine Karte hinein oder aus einer Karte heraus, um mehr Details bzw. mehr Überblick zu bekommen. Die kontinuierliche Generalisierung von Landkarten versucht die Änderungen zwischen verschiedenen Maßstäben stetig zu machen. Dies ist wichtig, um Nutzern eine angenehme Zoom-Erfahrung zu bieten. Um eine qualitativ hochwertige kontinuierliche Generalisierung zu erreichen, kann man wichtige Aspekte bei der Generierung von Online-Karten optimieren. In diesem Buch haben wir Optimierung bei der Generalisierung von Landnutzungskarten, von administrativen Grenzen, Gebäuden und Küstenlinien eingesetzt. Unsere Experimente zeigen, dass die kontinuierliche Generalisierung von Landkarten in der Tat von Optimierung profitiert. KW - land-cover area KW - administrative boundary KW - building KW - morphing KW - data structure KW - zooming KW - Generalisierung KW - Landnutzungskartierung KW - Optimierung Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-174427 SN - 978-3-95826-104-4 SN - 978-3-95826-105-1 N1 - Parallel erschienen als Druckausgabe in Würzburg University Press, 978-3-95826-104-4, 24,90 EUR. PB - Würzburg University Press CY - Würzburg ET - 1. Auflage ER - TY - JOUR A1 - Pfitzner, Christian A1 - May, Stefan A1 - Nüchter, Andreas T1 - Body weight estimation for dose-finding and health monitoring of lying, standing and walking patients based on RGB-D data JF - Sensors N2 - This paper describes the estimation of the body weight of a person in front of an RGB-D camera. A survey of different methods for body weight estimation based on depth sensors is given. First, an estimation of people standing in front of a camera is presented. Second, an approach based on a stream of depth images is used to obtain the body weight of a person walking towards a sensor. The algorithm first extracts features from a point cloud and forwards them to an artificial neural network (ANN) to obtain an estimation of body weight. Besides the algorithm for the estimation, this paper further presents an open-access dataset based on measurements from a trauma room in a hospital as well as data from visitors of a public event. In total, the dataset contains 439 measurements. The article illustrates the efficiency of the approach with experiments with persons lying down in a hospital, standing persons, and walking persons. Applicable scenarios for the presented algorithm are body weight-related dosing of emergency patients. KW - RGB-D KW - human body weight KW - image processing KW - kinect KW - machine learning KW - perception KW - segmentation KW - sensor fusion KW - stroke KW - thermal camera Y1 - 2018 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-176642 VL - 18 IS - 5 ER - TY - JOUR A1 - Sîrbu, Alina A1 - Becker, Martin A1 - Caminiti, Saverio A1 - De Baets, Bernard A1 - Elen, Bart A1 - Francis, Louise A1 - Gravino, Pietro A1 - Hotho, Andreas A1 - Ingarra, Stefano A1 - Loreto, Vittorio A1 - Molino, Andrea A1 - Mueller, Juergen A1 - Peters, Jan A1 - Ricchiuti, Ferdinando A1 - Saracino, Fabio A1 - Servedio, Vito D.P. A1 - Stumme, Gerd A1 - Theunis, Jan A1 - Tria, Francesca A1 - Van den Bossche, Joris T1 - Participatory Patterns in an International Air Quality Monitoring Initiative JF - PLoS ONE N2 - The issue of sustainability is at the top of the political and societal agenda, being considered of extreme importance and urgency. Human individual action impacts the environment both locally (e.g., local air/water quality, noise disturbance) and globally (e.g., climate change, resource use). Urban environments represent a crucial example, with an increasing realization that the most effective way of producing a change is involving the citizens themselves in monitoring campaigns (a citizen science bottom-up approach). This is possible by developing novel technologies and IT infrastructures enabling large citizen participation. Here, in the wider framework of one of the first such projects, we show results from an international competition where citizens were involved in mobile air pollution monitoring using low cost sensing devices, combined with a web-based game to monitor perceived levels of pollution. Measures of shift in perceptions over the course of the campaign are provided, together with insights into participatory patterns emerging from this study. Interesting effects related to inertia and to direct involvement in measurement activities rather than indirect information exposure are also highlighted, indicating that direct involvement can enhance learning and environmental awareness. In the future, this could result in better adoption of policies towards decreasing pollution. KW - transport microenvironments KW - exposure KW - pollution KW - carbon Y1 - 2015 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-151379 VL - 10 IS - 8 ER - TY - THES A1 - Fleszar, Krzysztof T1 - Network-Design Problems in Graphs and on the Plane T1 - Netzwerk-Design-Probleme in Graphen und auf der Ebene N2 - A network design problem defines an infinite set whose elements, called instances, describe relationships and network constraints. It asks for an algorithm that, given an instance of this set, designs a network that respects the given constraints and at the same time optimizes some given criterion. In my thesis, I develop algorithms whose solutions are optimum or close to an optimum value within some guaranteed bound. I also examine the computational complexity of these problems. Problems from two vast areas are considered: graphs and the Euclidean plane. In the Maximum Edge Disjoint Paths problem, we are given a graph and a subset of vertex pairs that are called terminal pairs. We are asked for a set of paths where the endpoints of each path form a terminal pair. The constraint is that any two paths share at most one inner vertex. The optimization criterion is to maximize the cardinality of the set. In the hard-capacitated k-Facility Location problem, we are given an integer k and a complete graph where the distances obey a given metric and where each node has two numerical values: a capacity and an opening cost. We are asked for a subset of k nodes, called facilities, and an assignment of all the nodes, called clients, to the facilities. The constraint is that the number of clients assigned to a facility cannot exceed the facility's capacity value. The optimization criterion is to minimize the total cost which consists of the total opening cost of the facilities and the total distance between the clients and the facilities they are assigned to. In the Stabbing problem, we are given a set of axis-aligned rectangles in the plane. We are asked for a set of horizontal line segments such that, for every rectangle, there is a line segment crossing its left and right edge. The optimization criterion is to minimize the total length of the line segments. In the k-Colored Non-Crossing Euclidean Steiner Forest problem, we are given an integer k and a finite set of points in the plane where each point has one of k colors. For every color, we are asked for a drawing that connects all the points of the same color. The constraint is that drawings of different colors are not allowed to cross each other. The optimization criterion is to minimize the total length of the drawings. In the Minimum Rectilinear Polygon for Given Angle Sequence problem, we are given an angle sequence of left (+90°) turns and right (-90°) turns. We are asked for an axis-parallel simple polygon where the angles of the vertices yield the given sequence when walking around the polygon in counter-clockwise manner. The optimization criteria considered are to minimize the perimeter, the area, and the size of the axis-parallel bounding box of the polygon. N2 - Ein Netzwerk-Design-Problem definiert eine unendliche Menge, deren Elemente, als Instanzen bezeichnet, Beziehungen und Beschränkungen in einem Netzwerk beschreiben. Die Lösung eines solchen Problems besteht aus einem Algorithmus, der auf die Eingabe einer beliebigen Instanz dieser Menge ein Netzwerk entwirft, welches die gegebenen Beschränkungen einhält und gleichzeitig ein gegebenes Kriterium optimiert. In meiner Dissertation habe ich Algorithmen entwickelt, deren Netzwerke stets optimal sind oder nachweisbar nahe am Optimum liegen. Zusätzlich habe ich die Berechnungskomplexität dieser Probleme untersucht. Dabei wurden Probleme aus zwei weiten Gebieten betrachtet: Graphen und der Euklidische Ebene. Im Maximum-Edge-Disjoint-Paths-Problem besteht die Eingabe aus einem Graphen und einer Teilmenge von Knotenpaaren, die wir mit Terminalpaare bezeichnen. Gesucht ist eine Menge von Pfaden, die Terminalpaare verbinden. Die Beschränkung ist, dass keine zwei Pfade einen gleichen inneren Knoten haben dürfen. Das Optimierungskriterium ist die Maximierung der Kardinalität dieser Menge. Im Hard-Capacitated-k-Facility-Location-Problem besteht die Eingabe aus einer Ganzzahl k und einem vollständigen Graphen, in welchem die Distanzen einer gegebenen Metrik unterliegen und in welchem jedem Knoten sowohl eine numerische Kapazität als auch ein Eröffnungskostenwert zugeschrieben ist. Gesucht ist eine Teilmenge von k Knoten, Facilities genannt, und eine Zuweisung aller Knoten, Clients genannt, zu den Facilities. Die Beschränkung ist, dass die Anzahl der Clients, die einer Facility zugewiesen sind, nicht deren Kapazität überschreiten darf. Das Optimierungskriterium ist die Minimierung der Gesamtkosten bestehend aus den Gesamteröffnungskosten der Facilities sowie der Gesamtdistanz zwischen den Clients und den ihnen zugewiesenen Facilities. Im Stabbing-Problem besteht die Eingabe aus einer Menge von achsenparallelen Rechtecken in der Ebene. Gesucht ist eine Menge von horizontalen Geradenstücken mit der Randbedingung, dass die linke und rechte Seite eines jeden Rechtecks von einem Geradenstück verbunden ist. Das Optimierungskriterium ist die Minimierung der Gesamtlänge aller Geradenstücke. Im k-Colored-Non-Crossing-Euclidean-Steiner-Forest-Problem besteht die Eingabe aus einer Ganzzahl k und einer endlichen Menge von Punkten in der Ebene, wobei jeder Punkt in einer von k Farben gefärbt ist. Gesucht ist für jede Farbe eine Zeichnung, in welcher alle Punkte der Farbe verbunden sind. Die Beschränkung ist, dass Zeichnungen verschiedener Farben sich nicht kreuzen dürfen. Das Optimierungskriterium ist die Minimierung des Gesamtintenverbrauchs, das heißt, der Gesamtlänge der Zeichnungen. Im Minimum-Rectilinear-Polygon-for-Given-Angle-Sequence-Problem besteht die Eingabe aus einer Folge von Links- (+90°) und Rechtsabbiegungen (-90°). Gesucht ist ein achsenparalleles Polygon dessen Eckpunkte die gegebene Folge ergeben, wenn man das Polygon gegen den Uhrzeigersinn entlangläuft. Die Optimierungskriterien sind die Minimierung des Umfangs und der inneren Fläche des Polygons sowie der Größe des notwendigen Zeichenblattes, d.h., des kleinsten Rechteckes, das das Polygon einschließt. N2 - Given points in the plane, connect them using minimum ink. Though the task seems simple, it turns out to be very time consuming. In fact, scientists believe that computers cannot efficiently solve it. So, do we have to resign? This book examines such NP-hard network-design problems, from connectivity problems in graphs to polygonal drawing problems on the plane. First, we observe why it is so hard to optimally solve these problems. Then, we go over to attack them anyway. We develop fast algorithms that find approximate solutions that are very close to the optimal ones. Hence, connecting points with slightly more ink is not hard. KW - Euklidische Ebene KW - Algorithmus KW - Komplexität KW - NP-schweres Problem KW - Graph KW - approximation algorithm KW - hardness KW - optimization KW - graphs KW - network KW - Optimierungsproblem KW - Approximationsalgorithmus KW - complexity KW - Euclidean plane Y1 - 2018 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-154904 SN - 978-3-95826-076-4 (Print) SN - 978-3-95826-077-1 (Online) N1 - Parallel erschienen als Druckausgabe in Würzburg University Press, ISBN 978-3-95826-076-4, 28,90 EUR. PB - Würzburg University Press CY - Würzburg ET - 1. Auflage ER - TY - THES A1 - Wojtkowiak, Harald T1 - Planungssystem zur Steigerung der Autonomie von Kleinstsatelliten T1 - Planningsystem to increase the autonomy of small satellites N2 - Der Betrieb von Satelliten wird sich in Zukunft gravierend ändern. Die bisher ausgeübte konventionelle Vorgehensweise, bei der die Planung der vom Satelliten auszuführenden Aktivitäten sowie die Kontrolle hierüber ausschließlich vom Boden aus erfolgen, stößt bei heutigen Anwendungen an ihre Grenzen. Im schlimmsten Fall verhindert dieser Umstand sogar die Erschließung bisher ungenutzter Möglichkeiten. Der Gewinn eines Satelliten, sei es in Form wissenschaftlicher Daten oder der Vermarktung satellitengestützter Dienste, wird daher nicht optimal ausgeschöpft. Die Ursache für dieses Problem lässt sich im Grunde auf eine ausschlaggebende Tatsache zurückführen: Konventionelle Satelliten können ihr Verhalten, d.h. die Folge ihrer Tätigkeiten, nicht eigenständig anpassen. Stattdessen erstellt das Bedienpersonal am Boden - vor allem die Operatoren - mit Hilfe von Planungssoftware feste Ablaufpläne, die dann in Form von Kommandosequenzen von den Bodenstationen aus an die jeweiligen Satelliten hochgeladen werden. Dort werden die Befehle lediglich überprüft, interpretiert und strikt ausgeführt. Die Abarbeitung erfolgt linear. Situationsbedingte Änderungen, wie sie vergleichsweise bei der Codeausführung von Softwareprogrammen durch Kontrollkonstrukte, zum Beispiel Schleifen und Verzweigungen, üblich sind, sind typischerweise nicht vorgesehen. Der Operator ist daher die einzige Instanz, die das Verhalten des Satelliten mittels Kommandierung, per Upload, beeinflussen kann, und auch nur dann, wenn ein direkter Funkkontakt zwischen Satellit und Bodenstation besteht. Die dadurch möglichen Reaktionszeiten des Satelliten liegen bestenfalls bei einigen Sekunden, falls er sich im Wirkungsbereich der Bodenstation befindet. Außerhalb des Kontaktfensters kann sich die Zeitschranke, gegeben durch den Orbit und die aktuelle Position des Satelliten, von einigen Minuten bis hin zu einigen Stunden erstrecken. Die Signallaufzeiten der Funkübertragung verlängern die Reaktionszeiten um weitere Sekunden im erdnahen Bereich. Im interplanetaren Raum erstrecken sich die Zeitspannen aufgrund der immensen Entfernungen sogar auf mehrere Minuten. Dadurch bedingt liegt die derzeit technologisch mögliche, bodengestützte, Reaktionszeit von Satelliten bestenfalls im Bereich von einigen Sekunden. Diese Einschränkung stellt ein schweres Hindernis für neuartige Satellitenmissionen, bei denen insbesondere nichtdeterministische und kurzzeitige Phänomene (z.B. Blitze und Meteoreintritte in die Erdatmosphäre) Gegenstand der Beobachtungen sind, dar. Die langen Reaktionszeiten des konventionellen Satellitenbetriebs verhindern die Realisierung solcher Missionen, da die verzögerte Reaktion erst erfolgt, nachdem das zu beobachtende Ereignis bereits abgeschlossen ist. Die vorliegende Dissertation zeigt eine Möglichkeit, das durch die langen Reaktionszeiten entstandene Problem zu lösen, auf. Im Zentrum des Lösungsansatzes steht dabei die Autonomie. Im Wesentlichen geht es dabei darum, den Satelliten mit der Fähigkeit auszustatten, sein Verhalten, d.h. die Folge seiner Tätigkeiten, eigenständig zu bestimmen bzw. zu ändern. Dadurch wird die direkte Abhängigkeit des Satelliten vom Operator bei Reaktionen aufgehoben. Im Grunde wird der Satellit in die Lage versetzt, sich selbst zu kommandieren. Die Idee der Autonomie wurde im Rahmen der zugrunde liegenden Forschungsarbeiten umgesetzt. Das Ergebnis ist ein autonomes Planungssystem. Dabei handelt es sich um ein Softwaresystem, mit dem sich autonomes Verhalten im Satelliten realisieren lässt. Es kann an unterschiedliche Satellitenmissionen angepasst werden. Ferner deckt es verschiedene Aspekte des autonomen Satellitenbetriebs, angefangen bei der generellen Entscheidungsfindung der Tätigkeiten, über die zeitliche Ablaufplanung unter Einbeziehung von Randbedingungen (z.B. Ressourcen) bis hin zur eigentlichen Ausführung, d.h. Kommandierung, ab. Das Planungssystem kommt als Anwendung in ASAP, einer autonomen Sensorplattform, zum Einsatz. Es ist ein optisches System und dient der Detektion von kurzzeitigen Phänomenen und Ereignissen in der Erdatmosphäre. Die Forschungsarbeiten an dem autonomen Planungssystem, an ASAP sowie an anderen zu diesen in Bezug stehenden Systemen wurden an der Professur für Raumfahrttechnik des Lehrstuhls Informatik VIII der Julius-Maximilians-Universität Würzburg durchgeführt. N2 - Satellite operation will change thoroughly in future. So far the currently performed conventional approach of controlling satellites is hitting its limitation by todays application. This is due to the fact that activities of the satellite are planned and controlled exclusively by ground infrastructure. In the worst case this circumstance prevents the exploitation of potential but so far unused opportunities. Thus the yield of satellites, may it be in the form of scientific research data or the commercialization of satellite services, is not optimized. After all the cause of this problem can be tracked back to one crucial matter: Conventional satellites are not able to alter their behaviour, i.e. the order of their actions, themselves. Instead schedules are created by ground staff – mainly operators - utilizing specialized planning software. The output is then transformed into command sequences and uploaded to the dedicated satellite via ground stations. On-board the commands are solely checked, interpreted and strictly executed. The flow is linear. Situational changes, like in the code execution of software programs via control constructs, e.g. loops and branches, are typically not present. Thus the operator is the only instance which is able to change the behaviour of the satellite via command upload. Therefore a direct radio link between satellite and ground station is required. Reaction times are hereby restricted. In the best case, means when the satellite is inside the area of effect, the limitations are to a few seconds. Outside the contact window, the time bound may increase from minutes to hours. The exact timing are dependant from the orbit of the satellite and its position on it. The signal flow of the radio links adds additional reaction time from a few seconds in near earth up to some minutes in interplanetary space due to the vast distances. In sum the best achievable ground based reaction time lies in the area of some seconds. This restriction is a severe handicap for novel satellite missions with focus on non-deterministic and short-time phenomena, e.g. lightning and meteor entries into Earth atmosphere. The long reaction times of conventional satellite operations prevent the realization of such missions. This is due to the fact that delayed reactions take place after the event to observe has finished. This dissertation shows a possibility to solve the problem caused by long reaction times. Autonomy lies in the centre of the main approach. The key is to augment the satellite with the ability to alter its behaviour, i.e. the sequence of its actions, and to deliberate about it. Thus, the direct reaction dependency of the satellite from operators will be lifted. In principle the satellite will be able to command itself. The herein idea of autonomy is based on research work, which provides the context for design and implementation. The output is an autonomous planning system. It’s a software system which enables a satellite to behave autonomously and can be adapted to different types of satellite missions. Additionally, it covers different aspects of autonomous satellite operation, starting with general decision making of activities, going over to time scheduling inclusive constraint consideration, e.g. resources, and finishing at last with the actual execution, i.e. commanding. The autonomous planning system runs as one application of ASAP, an autonomous sensor platform. It is an optical system with the purpose to detect short-time phenomena and events in Earth atmosphere. The research work for the autonomous planning system, for ASAP and for other related systems has been executed at the professorship for space technology which is part of the department for computer science VIII at the Julius-Maximilians-Universität Würzburg. KW - Planungssystem KW - Autonomie KW - Satellit KW - Entscheidungsfindung KW - Ablaufplanung KW - Planausführung KW - System KW - Missionsbetrieb KW - decission finding KW - scheduling KW - plan execution KW - system KW - mission operation Y1 - 2018 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-163569 ER - TY - JOUR A1 - Kaiser, Dennis A1 - Lesch, Veronika A1 - Rothe, Julian A1 - Strohmeier, Michael A1 - Spieß, Florian A1 - Krupitzer, Christian A1 - Montenegro, Sergio A1 - Kounev, Samuel T1 - Towards Self-Aware Multirotor Formations JF - Computers N2 - In the present day, unmanned aerial vehicles become seemingly more popular every year, but, without regulation of the increasing number of these vehicles, the air space could become chaotic and uncontrollable. In this work, a framework is proposed to combine self-aware computing with multirotor formations to address this problem. The self-awareness is envisioned to improve the dynamic behavior of multirotors. The formation scheme that is implemented is called platooning, which arranges vehicles in a string behind the lead vehicle and is proposed to bring order into chaotic air space. Since multirotors define a general category of unmanned aerial vehicles, the focus of this thesis are quadcopters, platforms with four rotors. A modification for the LRA-M self-awareness loop is proposed and named Platooning Awareness. The implemented framework is able to offer two flight modes that enable waypoint following and the self-awareness module to find a path through scenarios, where obstacles are present on the way, onto a goal position. The evaluation of this work shows that the proposed framework is able to use self-awareness to learn about its environment, avoid obstacles, and can successfully move a platoon of drones through multiple scenarios. KW - self-aware computing KW - unmanned aerial vehicles KW - multirotors KW - quadcopters KW - intelligent transportation systems Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-200572 SN - 2073-431X VL - 9 IS - 1 ER - TY - JOUR A1 - Grohmann, Johannes A1 - Herbst, Nikolas A1 - Chalbani, Avi A1 - Arian, Yair A1 - Peretz, Noam A1 - Kounev, Samuel T1 - A Taxonomy of Techniques for SLO Failure Prediction in Software Systems JF - Computers N2 - Failure prediction is an important aspect of self-aware computing systems. Therefore, a multitude of different approaches has been proposed in the literature over the past few years. In this work, we propose a taxonomy for organizing works focusing on the prediction of Service Level Objective (SLO) failures. Our taxonomy classifies related work along the dimensions of the prediction target (e.g., anomaly detection, performance prediction, or failure prediction), the time horizon (e.g., detection or prediction, online or offline application), and the applied modeling type (e.g., time series forecasting, machine learning, or queueing theory). The classification is derived based on a systematic mapping of relevant papers in the area. Additionally, we give an overview of different techniques in each sub-group and address remaining challenges in order to guide future research. KW - taxonomy KW - survey KW - failure prediction KW - anomaly prediction KW - anomaly detection KW - self-aware computing KW - self-adaptive systems KW - performance prediction Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-200594 SN - 2073-431X VL - 9 IS - 1 ER - TY - JOUR A1 - Oberdörfer, Sebastian A1 - Latoschik, Marc Erich T1 - Knowledge encoding in game mechanics: transfer-oriented knowledge learning in desktop-3D and VR JF - International Journal of Computer Games Technology N2 - Affine Transformations (ATs) are a complex and abstract learning content. Encoding the AT knowledge in Game Mechanics (GMs) achieves a repetitive knowledge application and audiovisual demonstration. Playing a serious game providing these GMs leads to motivating and effective knowledge learning. Using immersive Virtual Reality (VR) has the potential to even further increase the serious game’s learning outcome and learning quality. This paper compares the effectiveness and efficiency of desktop-3D and VR in respect to the achieved learning outcome. Also, the present study analyzes the effectiveness of an enhanced audiovisual knowledge encoding and the provision of a debriefing system. The results validate the effectiveness of the knowledge encoding in GMs to achieve knowledge learning. The study also indicates that VR is beneficial for the overall learning quality and that an enhanced audiovisual encoding has only a limited effect on the learning outcome. KW - games Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-201159 VL - 2019 ER - TY - THES A1 - Baier, Pablo A. T1 - Simulator for Minimally Invasive Vascular Interventions: Hardware and Software T1 - VR-Simulation für das Training von Herzkathetereingriffen: Hard- und Softwarelösung N2 - A complete simulation system is proposed that can be used as an educational tool by physicians in training basic skills of Minimally Invasive Vascular Interventions. In the first part, a surface model is developed to assemble arteries having a planar segmentation. It is based on Sweep Surfaces and can be extended to T- and Y-like bifurcations. A continuous force vector field is described, representing the interaction between the catheter and the surface. The computation time of the force field is almost unaffected when the resolution of the artery is increased. The mechanical properties of arteries play an essential role in the study of the circulatory system dynamics, which has been becoming increasingly important in the treatment of cardiovascular diseases. In Virtual Reality Simulators, it is crucial to have a tissue model that responds in real time. In this work, the arteries are discretized by a two dimensional mesh and the nodes are connected by three kinds of linear springs. Three tissue layers (Intima, Media, Adventitia) are considered and, starting from the stretch-energy density, some of the elasticity tensor components are calculated. The physical model linearizes and homogenizes the material response, but it still contemplates the geometric nonlinearity. In general, if the arterial stretch varies by 1% or less, then the agreement between the linear and nonlinear models is trustworthy. In the last part, the physical model of the wire proposed by Konings is improved. As a result, a simpler and more stable method is obtained to calculate the equilibrium configuration of the wire. In addition, a geometrical method is developed to perform relaxations. It is particularly useful when the wire is hindered in the physical method because of the boundary conditions. The physical and the geometrical methods are merged, resulting in efficient relaxations. Tests show that the shape of the virtual wire agrees with the experiment. The proposed algorithm allows real-time executions and the hardware to assemble the simulator has a low cost. N2 - Es wird ein vollständiges Simulationssystem entwickelt, das von Ärzten als Lehrmittel zur Ausbildung grundlegender Fertigkeiten bei Herzkathetereingriffen eingesetzt werden kann. Im ersten Teil wird ein Oberflächenmodell zur Erstellung von Arterien mit planarer Segmentierung entwickelt. Im zweiten Teil werden die Arterien durch ein zweidimensionales Netz diskretisiert, die Knoten werden durch drei Arten linearer Federn verbunden und ausgehend von einer Dehnungsenergie-Dichte-Funktion werden einige Komponenten des Elastizitätstensors berechnet. Im letzten Teil wird das von anderen Autoren vorgeschlagene physikalische Modell des Drahtes verbessert und eine neue geometrische Methode entwickelt. Der vorgeschlagene Algorithmus ermöglicht Echtzeit-Ausführungen. Die Hardware des Simulators hat geringe Herstellungskosten. T3 - Forschungsberichte in der Robotik = Research Notes in Robotics - 15 KW - Computersimulation KW - Simulator KW - Arterie KW - Elastizitätstensor KW - Herzkatheter KW - Minimally invasive vascular intervention KW - Wire relaxation KW - Artery KW - Elasticity tensor KW - Stiffness KW - educational tool KW - Elastizitätstensor KW - Herzkathetereingriff KW - Software KW - Hardware Y1 - 2018 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-161190 SN - 978-3-945459-22-5 ER - TY - CHAP ED - von Mammen, Sebastian ED - Klemke, Roland ED - Lorber, Martin T1 - Proceedings of the 1st Games Technology Summit BT - part of Clash of Realites 11th International Conference on the Technology and Theory of Digital Games N2 - As part of the Clash of Realities International Conference on the Technology and Theory of Digital Games, the Game Technology Summit is a premium venue to bring together experts from academia and industry to disseminate state-of-the-art research on trending technology topics in digital games. In this first iteration of the Game Technology Summit, we specifically paid attention on how the successes in AI in Natural User Interfaces have been impacting the games industry (industry track) and which scientific, state-of-the-art ideas and approaches are currently pursued (scientific track). KW - Veranstaltung KW - Künstliche Intelligenz KW - Mensch-Maschine-Kommunikation KW - Computerspiel KW - natural user interfaces KW - artificial intelligence Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-245776 SN - 978-3-945459-36-2 ER - TY - JOUR A1 - Du, Shitong A1 - Lauterbach, Helge A. A1 - Li, Xuyou A1 - Demisse, Girum G. A1 - Borrmann, Dorit A1 - Nüchter, Andreas T1 - Curvefusion — A Method for Combining Estimated Trajectories with Applications to SLAM and Time-Calibration JF - Sensors N2 - Mapping and localization of mobile robots in an unknown environment are essential for most high-level operations like autonomous navigation or exploration. This paper presents a novel approach for combining estimated trajectories, namely curvefusion. The robot used in the experiments is equipped with a horizontally mounted 2D profiler, a constantly spinning 3D laser scanner and a GPS module. The proposed algorithm first combines trajectories from different sensors to optimize poses of the planar three degrees of freedom (DoF) trajectory, which is then fed into continuous-time simultaneous localization and mapping (SLAM) to further improve the trajectory. While state-of-the-art multi-sensor fusion methods mainly focus on probabilistic methods, our approach instead adopts a deformation-based method to optimize poses. To this end, a similarity metric for curved shapes is introduced into the robotics community to fuse the estimated trajectories. Additionally, a shape-based point correspondence estimation method is applied to the multi-sensor time calibration. Experiments show that the proposed fusion method can achieve relatively better accuracy, even if the error of the trajectory before fusion is large, which demonstrates that our method can still maintain a certain degree of accuracy in an environment where typical pose estimation methods have poor performance. In addition, the proposed time-calibration method also achieves high accuracy in estimating point correspondences. KW - mapping KW - continuous-time SLAM KW - deformation-based method KW - time calibration Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-219988 SN - 1424-8220 VL - 20 IS - 23 ER - TY - JOUR A1 - Glémarec, Yann A1 - Lugrin, Jean-Luc A1 - Bosser, Anne-Gwenn A1 - Buche, Cédric A1 - Latoschik, Marc Erich T1 - Controlling the stage: a high-level control system for virtual audiences in Virtual Reality JF - Frontiers in Virtual Reality N2 - This article presents a novel method for controlling a virtual audience system (VAS) in Virtual Reality (VR) application, called STAGE, which has been originally designed for supervised public speaking training in university seminars dedicated to the preparation and delivery of scientific talks. We are interested in creating pedagogical narratives: narratives encompass affective phenomenon and rather than organizing events changing the course of a training scenario, pedagogical plans using our system focus on organizing the affects it arouses for the trainees. Efficiently controlling a virtual audience towards a specific training objective while evaluating the speaker’s performance presents a challenge for a seminar instructor: the high level of cognitive and physical demands required to be able to control the virtual audience, whilst evaluating speaker’s performance, adjusting and allowing it to quickly react to the user’s behaviors and interactions. It is indeed a critical limitation of a number of existing systems that they rely on a Wizard of Oz approach, where the tutor drives the audience in reaction to the user’s performance. We address this problem by integrating with a VAS a high-level control component for tutors, which allows using predefined audience behavior rules, defining custom ones, as well as intervening during run-time for finer control of the unfolding of the pedagogical plan. At its core, this component offers a tool to program, select, modify and monitor interactive training narratives using a high-level representation. The STAGE offers the following features: i) a high-level API to program pedagogical narratives focusing on a specific public speaking situation and training objectives, ii) an interactive visualization interface iii) computation and visualization of user metrics, iv) a semi-autonomous virtual audience composed of virtual spectators with automatic reactions to the speaker and surrounding spectators while following the pedagogical plan V) and the possibility for the instructor to embody a virtual spectator to ask questions or guide the speaker from within the Virtual Environment. We present here the design, and implementation of the tutoring system and its integration in STAGE, and discuss its reception by end-users. KW - virtual reality KW - virtual agent KW - behavior perception KW - public speaking KW - education Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-284601 SN - 2673-4192 VL - 3 ER - TY - THES A1 - Löffler, Andre T1 - Constrained Graph Layouts: Vertices on the Outer Face and on the Integer Grid T1 - Graphzeichnen unter Nebenbedingungen: Knoten auf der Außenfacette und mit ganzzahligen Koordinaten N2 - Constraining graph layouts - that is, restricting the placement of vertices and the routing of edges to obey certain constraints - is common practice in graph drawing. In this book, we discuss algorithmic results on two different restriction types: placing vertices on the outer face and on the integer grid. For the first type, we look into the outer k-planar and outer k-quasi-planar graphs, as well as giving a linear-time algorithm to recognize full and closed outer k-planar graphs Monadic Second-order Logic. For the second type, we consider the problem of transferring a given planar drawing onto the integer grid while perserving the original drawings topology; we also generalize a variant of Cauchy's rigidity theorem for orthogonal polyhedra of genus 0 to those of arbitrary genus. N2 - Das Einschränken von Zeichnungen von Graphen, sodass diese bestimmte Nebenbedingungen erfüllen - etwa solche, die das Platzieren von Knoten oder den Verlauf von Kanten beeinflussen - sind im Graphzeichnen allgegenwärtig. In dieser Arbeit befassen wir uns mit algorithmischen Resultaten zu zwei speziellen Einschränkungen, nämlich dem Platzieren von Knoten entweder auf der Außenfacette oder auf ganzzahligen Koordinaten. Für die erste Einschränkung untersuchen wir die außen k-planaren und außen k-quasi-planaren Graphen und geben einen auf monadische Prädikatenlogik zweiter Stufe basierenden Algorithmus an, der überprüft, ob ein Graph voll außen k-planar ist. Für die zweite Einschränkung untersuchen wir das Problem, eine gegebene planare Zeichnung eines Graphen auf das ganzzahlige Koordinatengitter zu transportieren, ohne dabei die Topologie der Zeichnung zu verändern; außerdem generalisieren wir eine Variante von Cauchys Starrheitssatz für orthogonale Polyeder von Geschlecht 0 auf solche von beliebigem Geschlecht. KW - Graphenzeichnen KW - Komplexität KW - Algorithmus KW - Algorithmische Geometrie KW - Kombinatorik KW - Planare Graphen KW - Polyeder KW - Konvexe Zeichnungen Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-215746 SN - 978-3-95826-146-4 SN - 978-3-95826-147-1 N1 - Parallel erschienen als Druckausgabe in Würzburg University Press, ISBN 978-3-95826-146-4, 32,90 EUR PB - Würzburg University Press CY - Würzburg ET - 1. Auflage ER - TY - JOUR A1 - Linsenmann, Thomas A1 - März, Alexander A1 - Dufner, Vera A1 - Stetter, Christian A1 - Weiland, Judith A1 - Westermaier, Thomas T1 - Optimization of radiation settings for angiography using 3D fluoroscopy for imaging of intracranial aneurysms JF - Computer Assisted Surgery N2 - Mobile 3D fluoroscopes have become increasingly available in neurosurgical operating rooms. We recently reported its use for imaging cerebral vascular malformations and aneurysms. This study was conducted to evaluate various radiation settings for the imaging of cerebral aneurysms before and after surgical occlusion. Eighteen patients with cerebral aneurysms with the indication for surgical clipping were included in this prospective analysis. Before surgery the patients were randomized into one of three different scan protocols according (default settings of the 3D fluoroscope): Group 1: 110 kV, 80 mA (enhanced cranial mode), group 2: 120 kV, 64 mA (lumbar spine mode), group 3: 120 kV, 25 mA (head/neck settings). Prior to surgery, a rotational fluoroscopy scan (duration 24 s) was performed without contrast agent followed by another scan with 50 ml of intravenous iodine contrast agent. The image files of both scans were transferred to an Apple PowerMac(R) workstation, subtracted and reconstructed using OsiriX(R) MD 10.0 software. The procedure was repeated after clip placement. The image quality regarding preoperative aneurysm configuration and postoperative assessment of aneurysm occlusion and vessel patency was analyzed by 2 independent reviewers using a 6-grade scale. This technique quickly supplies images of adequate quality to depict intracranial aneurysms and distal vessel patency after aneurysm clipping. Regarding these features, a further optimization to our previous protocol seems possible lowering the voltage and increasing tube current. For quick intraoperative assessment, image subtraction seems not necessary. Thus, a native scan without a contrast agent is not necessary. Further optimization may be possible using a different contrast injection protocol. KW - 3D fluoroscopy KW - aneurysm KW - fluoroscopy KW - intraoperative imaging Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-259251 VL - 26 IS - 1 ER - TY - RPRT A1 - Rossi, Angelo Pio A1 - Maurelli, Francesco A1 - Unnithan, Vikram A1 - Dreger, Hendrik A1 - Mathewos, Kedus A1 - Pradhan, Nayan A1 - Corbeanu, Dan-Andrei A1 - Pozzobon, Riccardo A1 - Massironi, Matteo A1 - Ferrari, Sabrina A1 - Pernechele, Claudia A1 - Paoletti, Lorenzo A1 - Simioni, Emanuele A1 - Maurizio, Pajola A1 - Santagata, Tommaso A1 - Borrmann, Dorit A1 - Nüchter, Andreas A1 - Bredenbeck, Anton A1 - Zevering, Jasper A1 - Arzberger, Fabian A1 - Reyes Mantilla, Camilo Andrés T1 - DAEDALUS - Descent And Exploration in Deep Autonomy of Lava Underground Structures BT - Open Space Innovation Platform (OSIP) Lunar Caves-System Study N2 - The DAEDALUS mission concept aims at exploring and characterising the entrance and initial part of Lunar lava tubes within a compact, tightly integrated spherical robotic device, with a complementary payload set and autonomous capabilities. The mission concept addresses specifically the identification and characterisation of potential resources for future ESA exploration, the local environment of the subsurface and its geologic and compositional structure. A sphere is ideally suited to protect sensors and scientific equipment in rough, uneven environments. It will house laser scanners, cameras and ancillary payloads. The sphere will be lowered into the skylight and will explore the entrance shaft, associated caverns and conduits. Lidar (light detection and ranging) systems produce 3D models with high spatial accuracy independent of lighting conditions and visible features. Hence this will be the primary exploration toolset within the sphere. The additional payload that can be accommodated in the robotic sphere consists of camera systems with panoramic lenses and scanners such as multi-wavelength or single-photon scanners. A moving mass will trigger movements. The tether for lowering the sphere will be used for data communication and powering the equipment during the descending phase. Furthermore, the connector tether-sphere will host a WIFI access point, such that data of the conduit can be transferred to the surface relay station. During the exploration phase, the robot will be disconnected from the cable, and will use wireless communication. Emergency autonomy software will ensure that in case of loss of communication, the robot will continue the nominal mission. T3 - Forschungsberichte in der Robotik = Research Notes in Robotics - 21 KW - Lunar Caves KW - Spherical Robot KW - Lunar Exploration KW - Mapping KW - 3D Laser Scanning KW - Mond KW - Daedalus-Projekt KW - Lava Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-227911 SN - 978-3-945459-33-1 SN - 1868-7466 ER - TY - THES A1 - Niebler, Thomas T1 - Extracting and Learning Semantics from Social Web Data T1 - Extraktion und Lernen von Semantik aus Social Web-Daten N2 - Making machines understand natural language is a dream of mankind that existed since a very long time. Early attempts at programming machines to converse with humans in a supposedly intelligent way with humans relied on phrase lists and simple keyword matching. However, such approaches cannot provide semantically adequate answers, as they do not consider the specific meaning of the conversation. Thus, if we want to enable machines to actually understand language, we need to be able to access semantically relevant background knowledge. For this, it is possible to query so-called ontologies, which are large networks containing knowledge about real-world entities and their semantic relations. However, creating such ontologies is a tedious task, as often extensive expert knowledge is required. Thus, we need to find ways to automatically construct and update ontologies that fit human intuition of semantics and semantic relations. More specifically, we need to determine semantic entities and find relations between them. While this is usually done on large corpora of unstructured text, previous work has shown that we can at least facilitate the first issue of extracting entities by considering special data such as tagging data or human navigational paths. Here, we do not need to detect the actual semantic entities, as they are already provided because of the way those data are collected. Thus we can mainly focus on the problem of assessing the degree of semantic relatedness between tags or web pages. However, there exist several issues which need to be overcome, if we want to approximate human intuition of semantic relatedness. For this, it is necessary to represent words and concepts in a way that allows easy and highly precise semantic characterization. This also largely depends on the quality of data from which these representations are constructed. In this thesis, we extract semantic information from both tagging data created by users of social tagging systems and human navigation data in different semantic-driven social web systems. Our main goal is to construct high quality and robust vector representations of words which can the be used to measure the relatedness of semantic concepts. First, we show that navigation in the social media systems Wikipedia and BibSonomy is driven by a semantic component. After this, we discuss and extend methods to model the semantic information in tagging data as low-dimensional vectors. Furthermore, we show that tagging pragmatics influences different facets of tagging semantics. We then investigate the usefulness of human navigational paths in several different settings on Wikipedia and BibSonomy for measuring semantic relatedness. Finally, we propose a metric-learning based algorithm in adapt pre-trained word embeddings to datasets containing human judgment of semantic relatedness. This work contributes to the field of studying semantic relatedness between words by proposing methods to extract semantic relatedness from web navigation, learn highquality and low-dimensional word representations from tagging data, and to learn semantic relatedness from any kind of vector representation by exploiting human feedback. Applications first and foremest lie in ontology learning for the Semantic Web, but also semantic search or query expansion. N2 - Einer der großen Träume der Menschheit ist es, Maschinen dazu zu bringen, natürliche Sprache zu verstehen. Frühe Versuche, Computer dahingehend zu programmieren, dass sie mit Menschen vermeintlich intelligente Konversationen führen können, basierten hauptsächlich auf Phrasensammlungen und einfachen Stichwortabgleichen. Solche Ansätze sind allerdings nicht in der Lage, inhaltlich adäquate Antworten zu liefern, da der tatsächliche Inhalt der Konversation nicht erfasst werden kann. Folgerichtig ist es notwendig, dass Maschinen auf semantisch relevantes Hintergrundwissen zugreifen können, um diesen Inhalt zu verstehen. Solches Wissen ist beispielsweise in Ontologien vorhanden. Ontologien sind große Datenbanken von vernetztem Wissen über Objekte und Gegenstände der echten Welt sowie über deren semantische Beziehungen. Das Erstellen solcher Ontologien ist eine sehr kostspielige und aufwändige Aufgabe, da oft tiefgreifendes Expertenwissen benötigt wird. Wir müssen also Wege finden, um Ontologien automatisch zu erstellen und aktuell zu halten, und zwar in einer Art und Weise, dass dies auch menschlichem Empfinden von Semantik und semantischer Ähnlichkeit entspricht. Genauer gesagt ist es notwendig, semantische Entitäten und deren Beziehungen zu bestimmen. Während solches Wissen üblicherweise aus Textkorpora extrahiert wird, ist es möglich, zumindest das erste Problem - semantische Entitäten zu bestimmen - durch Benutzung spezieller Datensätze zu umgehen, wie zum Beispiel Tagging- oder Navigationsdaten. In diesen Arten von Datensätzen ist es nicht notwendig, Entitäten zu extrahieren, da sie bereits aufgrund inhärenter Eigenschaften bei der Datenakquise vorhanden sind. Wir können uns also hauptsächlich auf die Bestimmung von semantischen Relationen und deren Intensität fokussieren. Trotzdem müssen hier noch einige Hindernisse überwunden werden. Beispielsweise ist es notwendig, Repräsentationen für semantische Entitäten zu finden, so dass es möglich ist, sie einfach und semantisch hochpräzise zu charakterisieren. Dies hängt allerdings auch erheblich von der Qualität der Daten ab, aus denen diese Repräsentationen konstruiert werden. In der vorliegenden Arbeit extrahieren wir semantische Informationen sowohl aus Taggingdaten, von Benutzern sozialer Taggingsysteme erzeugt, als auch aus Navigationsdaten von Benutzern semantikgetriebener Social Media-Systeme. Das Hauptziel dieser Arbeit ist es, hochqualitative und robuste Vektordarstellungen von Worten zu konstruieren, die dann dazu benutzt werden können, die semantische Ähnlichkeit von Konzepten zu bestimmen. Als erstes zeigen wir, dass Navigation in Social Media Systemen unter anderem durch eine semantische Komponente getrieben wird. Danach diskutieren und erweitern wir Methoden, um die semantische Information in Taggingdaten als niedrigdimensionale sogenannte “Embeddings” darzustellen. Darüberhinaus demonstrieren wir, dass die Taggingpragmatik verschiedene Facetten der Taggingsemantik beeinflusst. Anschließend untersuchen wir, inwieweit wir menschliche Navigationspfade zur Bestimmung semantischer Ähnlichkeit benutzen können. Hierzu betrachten wir mehrere Datensätze, die Navigationsdaten in verschiedenen Rahmenbedingungen beinhalten. Als letztes stellen wir einen neuartigen Algorithmus vor, um bereits trainierte Word Embeddings im Nachhinein an menschliche Intuition von Semantik anzupassen. Diese Arbeit steuert wertvolle Beiträge zum Gebiet der Bestimmung von semantischer Ähnlichkeit bei: Es werden Methoden vorgestellt werden, um hochqualitative semantische Information aus Web-Navigation und Taggingdaten zu extrahieren, diese mittels niedrigdimensionaler Vektordarstellungen zu modellieren und selbige schließlich besser an menschliches Empfinden von semantischer Ähnlichkeit anzupassen, indem aus genau diesem Empfinden gelernt wird. Anwendungen liegen in erster Linie darin, Ontologien für das Semantic Web zu lernen, allerdings auch in allen Bereichen, die Vektordarstellungen von semantischen Entitäten benutzen. KW - Semantik KW - Maschinelles Lernen KW - Soziale Software KW - Semantics KW - User Behavior KW - Social Web KW - Machine Learning Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-178666 ER - TY - JOUR A1 - Hein, Rebecca M. A1 - Latoschik, Marc Erich A1 - Wienrich, Carolin T1 - Inter- and transcultural learning in cocial virtual reality: a proposal for an inter- and transcultural virtual object database to be used in the implementation, reflection, and evaluation of virtual encounters JF - Multimodal Technologies and Interaction N2 - Visual stimuli are frequently used to improve memory, language learning or perception, and understanding of metacognitive processes. However, in virtual reality (VR), there are few systematically and empirically derived databases. This paper proposes the first collection of virtual objects based on empirical evaluation for inter-and transcultural encounters between English- and German-speaking learners. We used explicit and implicit measurement methods to identify cultural associations and the degree of stereotypical perception for each virtual stimuli (n = 293) through two online studies, including native German and English-speaking participants. The analysis resulted in a final well-describable database of 128 objects (called InteractionSuitcase). In future applications, the objects can be used as a great interaction or conversation asset and behavioral measurement tool in social VR applications, especially in the field of foreign language education. For example, encounters can use the objects to describe their culture, or teachers can intuitively assess stereotyped attitudes of the encounters. KW - virtual stimuli KW - implicit association test KW - virtual reality KW - social VR KW - InteractionSuitcase Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-278974 SN - 2414-4088 VL - 6 IS - 7 ER - TY - JOUR A1 - Djebko, Kirill A1 - Puppe, Frank A1 - Kayal, Hakan T1 - Model-based fault detection and diagnosis for spacecraft with an application for the SONATE triple cube nano-satellite JF - Aerospace N2 - The correct behavior of spacecraft components is the foundation of unhindered mission operation. However, no technical system is free of wear and degradation. A malfunction of one single component might significantly alter the behavior of the whole spacecraft and may even lead to a complete mission failure. Therefore, abnormal component behavior must be detected early in order to be able to perform counter measures. A dedicated fault detection system can be employed, as opposed to classical health monitoring, performed by human operators, to decrease the response time to a malfunction. In this paper, we present a generic model-based diagnosis system, which detects faults by analyzing the spacecraft’s housekeeping data. The observed behavior of the spacecraft components, given by the housekeeping data is compared to their expected behavior, obtained through simulation. Each discrepancy between the observed and the expected behavior of a component generates a so-called symptom. Given the symptoms, the diagnoses are derived by computing sets of components whose malfunction might cause the observed discrepancies. We demonstrate the applicability of the diagnosis system by using modified housekeeping data of the qualification model of an actual spacecraft and outline the advantages and drawbacks of our approach. KW - fault detection KW - model-based diagnosis KW - nano-satellite Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-198836 SN - 2226-4310 VL - 6 IS - 10 ER - TY - JOUR A1 - Zimmerer, Chris A1 - Fischbach, Martin A1 - Latoschik, Marc Erich T1 - Semantic Fusion for Natural Multimodal Interfaces using Concurrent Augmented Transition Networks JF - Multimodal Technologies and Interaction N2 - Semantic fusion is a central requirement of many multimodal interfaces. Procedural methods like finite-state transducers and augmented transition networks have proven to be beneficial to implement semantic fusion. They are compliant with rapid development cycles that are common for the development of user interfaces, in contrast to machine-learning approaches that require time-costly training and optimization. We identify seven fundamental requirements for the implementation of semantic fusion: Action derivation, continuous feedback, context-sensitivity, temporal relation support, access to the interaction context, as well as the support of chronologically unsorted and probabilistic input. A subsequent analysis reveals, however, that there is currently no solution for fulfilling the latter two requirements. As the main contribution of this article, we thus present the Concurrent Cursor concept to compensate these shortcomings. In addition, we showcase a reference implementation, the Concurrent Augmented Transition Network (cATN), that validates the concept’s feasibility in a series of proof of concept demonstrations as well as through a comparative benchmark. The cATN fulfills all identified requirements and fills the lack amongst previous solutions. It supports the rapid prototyping of multimodal interfaces by means of five concrete traits: Its declarative nature, the recursiveness of the underlying transition network, the network abstraction constructs of its description language, the utilized semantic queries, and an abstraction layer for lexical information. Our reference implementation was and is used in various student projects, theses, as well as master-level courses. It is openly available and showcases that non-experts can effectively implement multimodal interfaces, even for non-trivial applications in mixed and virtual reality. KW - multimodal fusion KW - multimodal interface KW - semantic fusion KW - procedural fusion methods KW - natural interfaces KW - human-computer interaction Y1 - 2018 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-197573 SN - 2414-4088 VL - 2 IS - 4 ER - TY - JOUR A1 - Loda, Sophia A1 - Krebs, Jonathan A1 - Danhof, Sophia A1 - Schreder, Martin A1 - Solimando, Antonio G. A1 - Strifler, Susanne A1 - Rasche, Leo A1 - Kortüm, Martin A1 - Kerscher, Alexander A1 - Knop, Stefan A1 - Puppe, Frank A1 - Einsele, Hermann A1 - Bittrich, Max T1 - Exploration of artificial intelligence use with ARIES in multiple myeloma research JF - Journal of Clinical Medicine N2 - Background: Natural language processing (NLP) is a powerful tool supporting the generation of Real-World Evidence (RWE). There is no NLP system that enables the extensive querying of parameters specific to multiple myeloma (MM) out of unstructured medical reports. We therefore created a MM-specific ontology to accelerate the information extraction (IE) out of unstructured text. Methods: Our MM ontology consists of extensive MM-specific and hierarchically structured attributes and values. We implemented “A Rule-based Information Extraction System” (ARIES) that uses this ontology. We evaluated ARIES on 200 randomly selected medical reports of patients diagnosed with MM. Results: Our system achieved a high F1-Score of 0.92 on the evaluation dataset with a precision of 0.87 and recall of 0.98. Conclusions: Our rule-based IE system enables the comprehensive querying of medical reports. The IE accelerates the extraction of data and enables clinicians to faster generate RWE on hematological issues. RWE helps clinicians to make decisions in an evidence-based manner. Our tool easily accelerates the integration of research evidence into everyday clinical practice. KW - natural language processing KW - ontology KW - artificial intelligence KW - multiple myeloma KW - real world evidence Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-197231 SN - 2077-0383 VL - 8 IS - 7 ER - TY - JOUR A1 - Lopez-Arreguin, A. J. R. A1 - Montenegro, S. T1 - Improving engineering models of terramechanics for planetary exploration JF - Results in Engineering N2 - This short letter proposes more consolidated explicit solutions for the forces and torques acting on typical rover wheels, that can be used as a method to determine their average mobility characteristics in planetary soils. The closed loop solutions stand in one of the verified methods, but at difference of the previous, observables are decoupled requiring a less amount of physical parameters to measure. As a result, we show that with knowledge of terrain properties, wheel driving performance rely in a single observable only. Because of their generality, the formulated equations established here can have further implications in autonomy and control of rovers or planetary soil characterization. KW - Wheel KW - Terramechanics KW - Forces KW - Torque KW - Robotics Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-202490 VL - 3 ER - TY - THES A1 - Budig, Benedikt T1 - Extracting Spatial Information from Historical Maps: Algorithms and Interaction T1 - Extraktion räumlicher Informationen aus historischen Landkarten: Algorithmen und Interaktion N2 - Historical maps are fascinating documents and a valuable source of information for scientists of various disciplines. Many of these maps are available as scanned bitmap images, but in order to make them searchable in useful ways, a structured representation of the contained information is desirable. This book deals with the extraction of spatial information from historical maps. This cannot be expected to be solved fully automatically (since it involves difficult semantics), but is also too tedious to be done manually at scale. The methodology used in this book combines the strengths of both computers and humans: it describes efficient algorithms to largely automate information extraction tasks and pairs these algorithms with smart user interactions to handle what is not understood by the algorithm. The effectiveness of this approach is shown for various kinds of spatial documents from the 16th to the early 20th century. N2 - Historische Landkarten sind faszinierende Dokumente und eine wertvolle Informationsquelle für Wissenschaftler verschiedener Fächer. Viele dieser Karten liegen als gescannte Bitmap-Bilder vor, aber um sie auf nützliche Art durchsuchbar zu machen ist eine strukturierte Repräsentation der enthaltenen Informationen wünschenswert. Dieses Buch beschäftigt sich mit der Extraktion räumlicher Informationen aus historischen Landkarten. Man kann nicht erwarten, dass dies vollautomatisch geschieht (da komplizierte Semantik involviert ist), aber es ist auch zu aufwändig, um im großen Stil manuell durchgeführt zu werden. Die Methodik, die in diesem Buch verwendet wird, kombiniert die Stärken von Computern und Menschen: Es werden effiziente Algorithmen beschrieben, die Extraktionsaufgaben weitgehend automatisieren, und dazu passende Nutzerinteraktionen entworfen, mit denen Fälle gelöst werden, die die Algorithmen nicht vestehen. Die Effekitivität dieses Ansatzes wird anhand verschiedener räumlicher Dokumente aus dem 16. bis frühen 20. Jahrhundert gezeigt. KW - Karte KW - Effizienter Algorithmus KW - Interaktion KW - Information Extraction KW - Smart User Interaction KW - Historical Maps KW - Itineraries KW - Deep Georeferencing KW - Benutzerinteraktion KW - Historische Landkarten KW - Itinerare KW - Georeferenzierung KW - Historische Karte KW - Raumdaten Y1 - 2018 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-160955 SN - 978-3-95826-092-4 SN - 978-3-95826-093-1 N1 - Parallel erschienen als Druckausgabe in Würzburg University Press, ISBN 978-3-95826-092-4, 32,90 Euro. PB - Würzburg University Press CY - Würzburg ET - 1. Auflage ER - TY - JOUR A1 - Rodrigues, Johannes A1 - Weiß, Martin A1 - Hewig, Johannes A1 - Allen, John J. B. T1 - EPOS: EEG Processing Open-Source Scripts JF - Frontiers in Neuroscience N2 - Background: Since the replication crisis, standardization has become even more important in psychological science and neuroscience. As a result, many methods are being reconsidered, and researchers’ degrees of freedom in these methods are being discussed as a potential source of inconsistencies across studies. New Method: With the aim of addressing these subjectivity issues, we have been working on a tutorial-like EEG (pre-)processing pipeline to achieve an automated method based on the semi-automated analysis proposed by Delorme and Makeig. Results: Two scripts are presented and explained step-by-step to perform basic, informed ERP and frequency-domain analyses, including data export to statistical programs and visual representations of the data. The open-source software EEGlab in MATLAB is used as the data handling platform, but scripts based on code provided by Mike Cohen (2014) are also included. Comparison with existing methods: This accompanying tutorial-like article explains and shows how the processing of our automated pipeline affects the data and addresses, especially beginners in EEG-analysis, as other (pre)-processing chains are mostly targeting rather informed users in specialized areas or only parts of a complete procedure. In this context, we compared our pipeline with a selection of existing approaches. Conclusion: The need for standardization and replication is evident, yet it is equally important to control the plausibility of the suggested solution by data exploration. Here, we provide the community with a tool to enhance the understanding and capability of EEG-analysis. We aim to contribute to comprehensive and reliable analyses for neuro-scientific research. KW - EEG KW - electroencephalography KW - event-related potentials-ERP KW - EEG processing KW - EEG preprocessing KW - EEG frequency band analysis Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-240221 SN - 1662-453X VL - 15 ER - TY - JOUR A1 - Hirth, Matthias A1 - Seufert, Michael A1 - Lange, Stanislav A1 - Meixner, Markus A1 - Tran-Gia, Phuoc T1 - Performance evaluation of hybrid crowdsensing and fixed sensor systems for event detection in urban environments JF - Sensors N2 - Crowdsensing offers a cost-effective way to collect large amounts of environmental sensor data; however, the spatial distribution of crowdsensing sensors can hardly be influenced, as the participants carry the sensors, and, additionally, the quality of the crowdsensed data can vary significantly. Hybrid systems that use mobile users in conjunction with fixed sensors might help to overcome these limitations, as such systems allow assessing the quality of the submitted crowdsensed data and provide sensor values where no crowdsensing data are typically available. In this work, we first used a simulation study to analyze a simple crowdsensing system concerning the detection performance of spatial events to highlight the potential and limitations of a pure crowdsourcing system. The results indicate that even if only a small share of inhabitants participate in crowdsensing, events that have locations correlated with the population density can be easily and quickly detected using such a system. On the contrary, events with uniformly randomly distributed locations are much harder to detect using a simple crowdsensing-based approach. A second evaluation shows that hybrid systems improve the detection probability and time. Finally, we illustrate how to compute the minimum number of fixed sensors for the given detection time thresholds in our exemplary scenario. KW - crowdsensing KW - event detection KW - detection time simulation KW - performance analysis Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-245245 SN - 1424-8220 VL - 21 IS - 17 ER - TY - INPR A1 - Dandekar, Thomas T1 - Analysing the phase space of the standard model and its basic four forces from a qubit phase transition perspective: implications for large-scale structure generation and early cosmological events N2 - The phase space for the standard model of the basic four forces for n quanta includes all possible ensemble combinations of their quantum states m, a total of n**m states. Neighbor states reach according to transition possibilities (S-matrix) with emergent time from entropic ensemble gradients. We replace the “big bang” by a condensation event (interacting qubits become decoherent) and inflation by a crystallization event – the crystal unit cell guarantees same symmetries everywhere. Interacting qubits solidify and form a rapidly growing domain where the n**m states become separated ensemble states, rising long-range forces stop ultimately further growth. After that very early events, standard cosmology with the hot fireball model takes over. Our theory agrees well with lack of inflation traces in cosmic background measurements, large-scale structure of voids and filaments, supercluster formation, galaxy formation, dominance of matter and life-friendliness. We prove qubit interactions to be 1,2,4 or 8 dimensional (agrees with E8 symmetry of our universe). Repulsive forces at ultrashort distances result from quantization, long-range forces limit crystal growth. Crystals come and go in the qubit ocean. This selects for the ability to lay seeds for new crystals, for self-organization and life-friendliness. We give energy estimates for free qubits vs bound qubits, misplacements in the qubit crystal and entropy increase during qubit decoherence / crystal formation. Scalar fields for color interaction and gravity derive from the permeating qubit-interaction field. Hence, vacuum energy gets low only inside the qubit crystal. Condensed mathematics may advantageously model free / bound qubits in phase space. KW - phase space KW - cosmology KW - emergent time KW - qubit KW - phase transition KW - bit Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-298580 ER - TY - RPRT A1 - Lhamo, Osel A1 - Nguyen, Giang T. A1 - Fitzek, Frank H. P. T1 - Virtual Queues for QoS Compliance of Haptic Data Streams in Teleoperation T2 - Würzburg Workshop on Next-Generation Communication Networks (WueWoWas'22) N2 - Tactile Internet aims at allowing perceived real-time interactions between humans and machines. This requires satisfying a stringent latency requirement of haptic data streams whose data rates vary drastically as the results of perceptual codecs. This introduces a complex problem for the underlying network infrastructure to fulfill the pre-defined level of Quality of Service (QoS). However, novel networking hardware with data plane programming capability allows processing packets differently and opens up a new opportunity. For example, a dynamic and network-aware resource management strategy can help satisfy the QoS requirements of different priority flows without wasting precious bandwidth. This paper introduces virtual queues for service differentiation between different types of traffic streams, leveraging protocol independent switch architecture (PISA). We propose coordinating the management of all the queues and dynamically adapting their sizes to minimize packet loss and delay due to network congestion and ensure QoS compliance. KW - Datennetz KW - data plane programming KW - software defined network KW - P4 KW - virtual queue KW - haptic data Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-280762 ER - TY - RPRT A1 - Vomhoff, Viktoria A1 - Geißler, Stefan A1 - Hoßfeld, Tobias T1 - Identification of Signaling Patterns in Mobile IoT Signaling Traffic T2 - Würzburg Workshop on Next-Generation Communication Networks (WueWoWas'22) N2 - We attempt to identify sequences of signaling dialogs, to strengthen our understanding of the signaling behavior of IoT devices by examining a dataset containing over 270.000 distinct IoT devices whose signaling traffic has been observed over a 31-day period in a 2G network [4]. We propose a set of rules that allows the assembly of signaling dialogs into so-called sessions in order to identify common patterns and lay the foundation for future research in the areas of traffic modeling and anomaly detection. KW - Datennetz KW - IoT Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-280819 ER - TY - GEN A1 - Funken, Matthias A1 - Tscherner, Michael T1 - Jahresbericht 2018 des Rechenzentrums der Universität Würzburg T1 - Annual Report 2018 of the Computer Center, University of Wuerzburg N2 - Eine Übersicht über die Aktivitäten des Rechenzentrums im Jahr 2018. T3 - Jahresbericht des Rechenzentrums der Universität Würzburg - 2018 KW - Julius-Maximilians-Universität Würzburg KW - Jahresbericht KW - Jahresbericht KW - Rechenzentrum KW - RZUW KW - annual report KW - Computer Center University of Wuerzburg Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-188265 UR - https://www.rz.uni-wuerzburg.de/wir/publikationen/ ET - 1. Auflage ER - TY - GEN T1 - Jahresbericht 2016 des Rechenzentrums der Universität Würzburg T1 - Annual Report 2016 of the Computer Center, University of Wuerzburg N2 - Das Dokument umfasst eine jährliche Zusammenfassung der Aktivitäten des Rechenzentrums als zentraler IT-Dienstleister der Universität Würzburg T3 - Jahresbericht des Rechenzentrums der Universität Würzburg - 2016 KW - Jahresbericht KW - Julius-Maximilians-Universität Würzburg KW - Rechenzentrum KW - annual report KW - Computer Center University of Wuerzburg KW - RZUW Y1 - 2017 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-153558 UR - https://www.rz.uni-wuerzburg.de/wir/publikationen/ ET - 1. Auflage ER - TY - GEN A1 - Funken, Matthias A1 - Tscherner, Michael T1 - Jahresbericht 2017 des Rechenzentrums der Universität Würzburg T1 - Annual Report 2017 of the Computer Center, University of Wuerzburg N2 - Eine Übersicht über die Aktivitäten des Rechenzentrums im Jahr 2017. T3 - Jahresbericht des Rechenzentrums der Universität Würzburg - 2017 KW - Julius-Maximilians-Universität Würzburg KW - RZUW KW - Jahresbericht KW - Rechenzentrum KW - Computer Center University of Wuerzburg KW - annual report Y1 - 2018 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-168537 UR - https://www.rz.uni-wuerzburg.de/wir/publikationen/ ET - 1. Auflage ER - TY - JOUR A1 - Scherer, Marc A1 - Fleishman, Sarel J. A1 - Jones, Patrik R. A1 - Dandekar, Thomas A1 - Bencurova, Elena T1 - Computational Enzyme Engineering Pipelines for Optimized Production of Renewable Chemicals JF - Frontiers in Bioengineering and Biotechnology N2 - To enable a sustainable supply of chemicals, novel biotechnological solutions are required that replace the reliance on fossil resources. One potential solution is to utilize tailored biosynthetic modules for the metabolic conversion of CO2 or organic waste to chemicals and fuel by microorganisms. Currently, it is challenging to commercialize biotechnological processes for renewable chemical biomanufacturing because of a lack of highly active and specific biocatalysts. As experimental methods to engineer biocatalysts are time- and cost-intensive, it is important to establish efficient and reliable computational tools that can speed up the identification or optimization of selective, highly active, and stable enzyme variants for utilization in the biotechnological industry. Here, we review and suggest combinations of effective state-of-the-art software and online tools available for computational enzyme engineering pipelines to optimize metabolic pathways for the biosynthesis of renewable chemicals. Using examples relevant for biotechnology, we explain the underlying principles of enzyme engineering and design and illuminate future directions for automated optimization of biocatalysts for the assembly of synthetic metabolic pathways. KW - computational KW - enzyme KW - engineering KW - design KW - biomanufacturing KW - biofuel KW - microbes KW - metabolism Y1 - 2021 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-240598 SN - 2296-4185 VL - 9 ER -