TY - JOUR A1 - Klemz, Boris A1 - Rote, Günter T1 - Linear-Time Algorithms for Maximum-Weight Induced Matchings and Minimum Chain Covers in Convex Bipartite Graphs JF - Algorithmica N2 - A bipartite graph G=(U,V,E) is convex if the vertices in V can be linearly ordered such that for each vertex u∈U, the neighbors of u are consecutive in the ordering of V. An induced matching H of G is a matching for which no edge of E connects endpoints of two different edges of H. We show that in a convex bipartite graph with n vertices and m weighted edges, an induced matching of maximum total weight can be computed in O(n+m) time. An unweighted convex bipartite graph has a representation of size O(n) that records for each vertex u∈U the first and last neighbor in the ordering of V. Given such a compact representation, we compute an induced matching of maximum cardinality in O(n) time. In convex bipartite graphs, maximum-cardinality induced matchings are dual to minimum chain covers. A chain cover is a covering of the edge set by chain subgraphs, that is, subgraphs that do not contain induced matchings of more than one edge. Given a compact representation, we compute a representation of a minimum chain cover in O(n) time. If no compact representation is given, the cover can be computed in O(n+m) time. All of our algorithms achieve optimal linear running time for the respective problem and model, and they improve and generalize the previous results in several ways: The best algorithms for the unweighted problem versions had a running time of O(n\(^{2}\)) (Brandstädt et al. in Theor. Comput. Sci. 381(1–3):260–265, 2007. https://doi.org/10.1016/j.tcs.2007.04.006). The weighted case has not been considered before. KW - dynamic programming KW - graph algorithm KW - induced matching KW - chain cover KW - convex bipartite graph KW - certifying algorithm Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-267876 SN - 1432-0541 VL - 84 IS - 4 ER - TY - JOUR A1 - Fischer, Norbert A1 - Hartelt, Alexander A1 - Puppe, Frank T1 - Line-level layout recognition of historical documents with background knowledge JF - Algorithms N2 - Digitization and transcription of historic documents offer new research opportunities for humanists and are the topics of many edition projects. However, manual work is still required for the main phases of layout recognition and the subsequent optical character recognition (OCR) of early printed documents. This paper describes and evaluates how deep learning approaches recognize text lines and can be extended to layout recognition using background knowledge. The evaluation was performed on five corpora of early prints from the 15th and 16th Centuries, representing a variety of layout features. While the main text with standard layouts could be recognized in the correct reading order with a precision and recall of up to 99.9%, also complex layouts were recognized at a rate as high as 90% by using background knowledge, the full potential of which was revealed if many pages of the same source were transcribed. KW - layout recognition KW - background knowledge KW - historical document analysis KW - fully convolutional neural networks KW - baseline detection KW - text line detection Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-310938 SN - 1999-4893 VL - 16 IS - 3 ER - TY - THES A1 - Somody, Joseph Christian Campbell T1 - Leveraging deep learning for identification and structural determination of novel protein complexes from \(in\) \(situ\) electron cryotomography of \(Mycoplasma\) \(pneumoniae\) T1 - Tiefenlernen als Werkzeug zur Identifizierung und Strukturbestimmung neuer Proteinkomplexe aus der \(in\)-\(situ\)-Elektronenkryotomographie von \(Mycoplasma\) \(pneumoniae\) N2 - The holy grail of structural biology is to study a protein in situ, and this goal has been fast approaching since the resolution revolution and the achievement of atomic resolution. A cell's interior is not a dilute environment, and proteins have evolved to fold and function as needed in that environment; as such, an investigation of a cellular component should ideally include the full complexity of the cellular environment. Imaging whole cells in three dimensions using electron cryotomography is the best method to accomplish this goal, but it comes with a limitation on sample thickness and produces noisy data unamenable to direct analysis. This thesis establishes a novel workflow to systematically analyse whole-cell electron cryotomography data in three dimensions and to find and identify instances of protein complexes in the data to set up a determination of their structure and identity for success. Mycoplasma pneumoniae is a very small parasitic bacterium with fewer than 700 protein-coding genes, is thin enough and small enough to be imaged in large quantities by electron cryotomography, and can grow directly on the grids used for imaging, making it ideal for exploratory studies in structural proteomics. As part of the workflow, a methodology for training deep-learning-based particle-picking models is established. As a proof of principle, a dataset of whole-cell Mycoplasma pneumoniae tomograms is used with this workflow to characterize a novel membrane-associated complex observed in the data. Ultimately, 25431 such particles are picked from 353 tomograms and refined to a density map with a resolution of 11 Å. Making good use of orthogonal datasets to filter search space and verify results, structures were predicted for candidate proteins and checked for suitable fit in the density map. In the end, with this approach, nine proteins were found to be part of the complex, which appears to be associated with chaperone activity and interact with translocon machinery. Visual proteomics refers to the ultimate potential of in situ electron cryotomography: the comprehensive interpretation of tomograms. The workflow presented here is demonstrated to help in reaching that potential. N2 - Der heilige Gral der Strukturbiologie ist die Untersuchung eines Proteins in situ, und dieses Ziel ist seit der Auflösungsrevolution und dem Erreichen der atomaren Auflösung in greifbare Nähe gerückt. Das Innere einer Zelle ist keine verdünnte Umgebung, und Proteine haben sich so entwickelt, dass sie sich falten und so funktionieren, wie es in dieser Umgebung erforderlich ist; daher sollte die Untersuchung einer zellulären Komponente idealerweise die gesamte Komplexität der zellulären Umgebung umfassen. Die Abbildung ganzer Zellen in drei Dimensionen mit Hilfe der Elektronenkryotomographie ist die beste Methode, um dieses Ziel zu erreichen, aber sie ist mit einer Beschränkung der Probendicke verbunden und erzeugt verrauschte Daten, die sich nicht für eine direkte Analyse eignen. In dieser Dissertation wird ein neuartiger Workflow zur systematischen dreidimensionalen Analyse von Ganzzell-Elektronenkryotomographiedaten und zur Auffindung und Identifizierung von Proteinkomplexen in diesen Daten entwickelt, um eine erfolgreiche Bestimmung ihrer Struktur und Identität zu ermöglichen. Mycoplasma pneumoniae ist ein sehr kleines parasitäres Bakterium mit weniger als 700 proteinkodierenden Genen. Es ist dünn und klein genug, um in grossen Mengen durch Elektronenkryotomographie abgebildet zu werden, und kann direkt auf den für die Abbildung verwendeten Gittern wachsen, was es ideal für Sondierungsstudien in der strukturellen Proteomik macht. Als Teil des Workflows wird eine Methodik für das Training von Deep-Learning-basierten Partikelpicken-Modellen entwickelt. Als Proof-of-Principle wird ein Dataset von Ganzzell-Tomogrammen von Mycoplasma pneumoniae mit diesem Workflow verwendet, um einen neuartigen membranassoziierten Komplex zu charakterisieren, der in den Daten beobachtet wurde. Insgesamt wurden 25431 solcher Partikel aus 353 Tomogrammen gepickt und zu einer Dichtekarte mit einer Auflösung von 11 Å verfeinert. Unter Verwendung orthogonaler Datensätze zur Filterung des Suchraums und zur Überprüfung der Ergebnisse wurden Strukturen für Protein-Kandidaten vorhergesagt und auf ihre Eignung für die Dichtekarte überprüft. Letztendlich wurden mit diesem Ansatz neun Proteine als Bestandteile des Komplexes gefunden, der offenbar mit der Chaperonaktivität in Verbindung steht und mit der Translocon-Maschinerie interagiert. Das ultimative Potenzial der In-situ-Elektronenkryotomographie – die umfassende Interpretation von Tomogrammen – wird als visuelle Proteomik bezeichnet. Der hier vorgestellte Workflow soll dabei helfen, dieses Potenzial auszuschöpfen. KW - Kryoelektronenmikroskopie KW - Tomografie KW - Mycoplasma pneumoniae KW - Deep learning KW - cryo-EM KW - cryo-ET KW - tomography KW - mycoplasma KW - pneumoniae KW - deep learning KW - particle picking KW - membrane protein KW - visual proteomics Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-313447 ER - TY - THES A1 - Dang, Nghia Duc T1 - Konzeption und Evaluation eines hybriden, skalierbaren Werkzeugs zur mechatronischen Systemdiagnose am Beispiel eines Diagnosesystems für freie Kfz-Werkstätten T1 - Conception and evaluation of a hybrid, scalable tool for mechatronical system diagnosis on the example of a diagnostic system for independent car repair shops N2 - Die Entwicklung eines wissensbasierten Systems, speziell eines Diagnosesystems, ist eine Teildisziplin der künstlichen Intelligenz und angewandten Informatik. Im Laufe der Forschung auf diesem Gebiet wurden verschiedene Lösungsansätze mit unterschiedlichem Erfolg bei der Anwendung in der Kraftfahrzeugdiagnose entwickelt. Diagnosesysteme in Vertragswerkstätten, das heißt in Fahrzeughersteller gebundenen Werkstätten, wenden hauptsächlich die fallbasierte Diagnostik an. Zum einen hält sich hier die Fahrzeugvielfalt in Grenzen und zum anderen besteht eine Meldepflicht bei neuen, nicht im System vorhandenen Fällen. Die freien Werkstätten verfügen nicht über eine solche Datenbank. Somit ist der fallbasierte Ansatz schwer umsetzbar. In freien Werkstätten - Fahrzeughersteller unabhängigen Werkstätten - basiert die Fehlersuche hauptsächlich auf Fehlerbäumen. Wegen der wachsenden Fahrzeugkomplexität, welche wesentlich durch die stark zunehmende Anzahl der durch mechatronische Systeme realisierten Funktionen bedingt ist, und der steigenden Typenvielfalt ist die geführte Fehlersuche in freien Werkstätten nicht immer zielführend. Um die Unterstützung des Personals von freien Werkstätten bei der zukünftigen Fehlersuche zu gewährleisten, werden neue Generationen von herstellerunabhängigen Diagnosetools benötigt, die die Probleme der Variantenvielfalt und Komplexität lösen. In der vorliegenden Arbeit wird ein Lösungsansatz vorgestellt, der einen qualitativen, modellbasierten Diagnoseansatz mit einem auf heuristischem Diagnosewissen basierenden Ansatz vereint. Neben der Grundlage zur Wissenserhebung werden in dieser Arbeit die theoretische Grundlage zur Beherrschung der Variantenvielfalt sowie die Tests für die erstellten Diagnosemodelle behandelt. Die Diagnose ist symptombasiert und die Inferenzmechanismen zur Verarbeitung des Diagnosewissens sind eine Kombination aus Propagierung der abweichenden physikalischen Größen im Modell und der Auswertung des heuristischen Wissens. Des Weiteren werden in dieser Arbeit verschiedene Aspekte der Realisierung der entwickelten theoretischen Grundlagen dargestellt, zum Beispiel: Systemarchitektur, Wissenserhebungsprozess, Ablauf des Diagnosevorgangs in den Werkstätten. Die Evaluierung der entwickelten Lösung bei der Wissenserhebung in Form von Modellerstellungen und Modellierungsworkshops sowie Feldtests dient nicht nur zur Bestätigung des entwickelten Ansatzes, sondern auch zur Ideenfindung für die Integration der entwickelten Tools in die existierende IT-Infrastruktur. N2 - The development of a knowledge-based system - in particular a diagnostic system - is a branch of artificial intelligence and applied computer science. Throughout the research in this field, various approaches have been developed with varying degrees of success in the application in automotive diagnostics. Diagnostic systems in authorized garages i.e. in garages bound to vehicle manufacturers mainly apply case-based diagnosis. In those cases first the variance of vehicles is limited and second there is a reporting obligation of new, yet not existing cases within the system. The independent repair shops do not have such a database. Thus, the case-based approach is difficult to implement. In independent garages - car manufacturer independent garages - the troubleshooting is mainly based on fault trees. Because of the growing complexity of vehicles, which is mainly due to the rapidly increasing number of functions realized by mechatronic systems and the increasing variety of vehicle types, guided troubleshooting for independent garages is not always productive. To ensure the support of the staff of independent garages in the future troubleshooting, new generations of multivendor diagnostic tools are needed to solve the problems of the variety and complexity. In the present paper an approach is presented which combines a qualitative, model-based diagnostic approach to a heuristic diagnostic knowledge-based approach. In addition to the basis for knowledge collection in this study the theoretical basis for the control of the variety and the tests for the generated diagnosis models are discussed. The diagnosis is symptom-based and the inference mechanisms for processing of diagnostic knowledge are a combination of propagation of the different physical parameters in the model and the evaluation of heuristic knowledge. Furthermore, various aspects of the implementation of the developed theoretical basis are presented in this thesis, for example: system architecture, knowledge collection process, the course of action of the diagnostic process in the workshops. The evaluation of the developed solution for the collection of knowledge in the form of preparation of models and modeling workshops, as well as field tests not only serves to validate the developed approach, but also to find ideas for the integration of the developed tools into the existing IT infrastructure. KW - Diagnosesystem KW - Künstliche Intelligenz KW - Modellbasierte Diagnose KW - Werkstattdiagnose KW - hybride Diagnose KW - skalierbare Diagnose KW - Inferenz KW - Angewandte Informatik KW - aftermarket diagnostic KW - model-base diagnosis KW - hybrid Diagnostic Y1 - 2012 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-70774 ER - TY - THES A1 - Atzmüller, Martin T1 - Knowledge-Intensive Subgroup Mining - Techniques for Automatic and Interactive Discovery T1 - Wissensintensive Subgruppenentdeckung – Automatische und Interaktive Entdeckungsmethoden N2 - Data mining has proved its significance in various domains and applications. As an important subfield of the general data mining task, subgroup mining can be used, e.g., for marketing purposes in business domains, or for quality profiling and analysis in medical domains. The goal is to efficiently discover novel, potentially useful and ultimately interesting knowledge. However, in real-world situations these requirements often cannot be fulfilled, e.g., if the applied methods do not scale for large data sets, if too many results are presented to the user, or if many of the discovered patterns are already known to the user. This thesis proposes a combination of several techniques in order to cope with the sketched problems: We discuss automatic methods, including heuristic and exhaustive approaches, and especially present the novel SD-Map algorithm for exhaustive subgroup discovery that is fast and effective. For an interactive approach we describe techniques for subgroup introspection and analysis, and we present advanced visualization methods, e.g., the zoomtable that directly shows the most important parameters of a subgroup and that can be used for optimization and exploration. We also describe various visualizations for subgroup comparison and evaluation in order to support the user during these essential steps. Furthermore, we propose to include possibly available background knowledge that is easy to formalize into the mining process. We can utilize the knowledge in many ways: To focus the search process, to restrict the search space, and ultimately to increase the efficiency of the discovery method. We especially present background knowledge to be applied for filtering the elements of the problem domain, for constructing abstractions, for aggregating values of attributes, and for the post-processing of the discovered set of patterns. Finally, the techniques are combined into a knowledge-intensive process supporting both automatic and interactive methods for subgroup mining. The practical significance of the proposed approach strongly depends on the available tools. We introduce the VIKAMINE system as a highly-integrated environment for knowledge-intensive active subgroup mining. Also, we present an evaluation consisting of two parts: With respect to objective evaluation criteria, i.e., comparing the efficiency and the effectiveness of the subgroup discovery methods, we provide an experimental evaluation using generated data. For that task we present a novel data generator that allows a simple and intuitive specification of the data characteristics. The results of the experimental evaluation indicate that the novel SD-Map method outperforms the other described algorithms using data sets similar to the intended application concerning the efficiency, and also with respect to precision and recall for the heuristic methods. Subjective evaluation criteria include the user acceptance, the benefit of the approach, and the interestingness of the results. We present five case studies utilizing the presented techniques: The approach has been successfully implemented in medical and technical applications using real-world data sets. The method was very well accepted by the users that were able to discover novel, useful, and interesting knowledge. N2 - Data Mining wird mit großem Erfolg in vielen Domänen angewandt. Subgruppenentdeckung als wichtiges Teilgebiet des Data Mining kann zum Beispiel gut im Marketing, oder zur Qualitätskontrolle und Analyse in medizinischen Domänen eingesetzt werden. Das allgemeine Ziel besteht darin, potentiell nützliches and letztendlich interessantes Wissen zu entdecken. Jedoch können diese Anforderungen im praktischen Einsatz oft nicht erfüllt werden, etwa falls die eingesetzten Methoden eine schlechte Skalierbarkeit für größere Datensätze aufweisen, falls dem Benutzer zu viele Ergebnisse präsentiert werden, oder falls der Anwender viele der gefundenen Subgruppen-Muster schon kennt. Diese Arbeit stellt eine Kombination von automatischen und interaktiven Techniken vor, um mit den genannten Problemen besser umgehen zu können: Es werden automatische heuristische und vollständige Subgruppenentdeckungs-Verfahren diskutiert, und insbesondere der neuartige SD-Map Algorithmus zur vollständigen Subgruppenentdeckung vorgestellt der sowohl schnell als auch effektiv ist. Bezüglich der interaktiven Techniken werden Methoden zur Subgruppen-Introspektion und Analyse, und fortgeschrittene Visualisierungstechniken vorgestellt, beispielsweise die Zoomtable, die die für die Subgruppenentdeckung wichtigsten Parameter direkt visualisiert und zur Optimierung und Exploration eingesetzt werden kann. Zusätzlich werden verschiedene Visualisierungen zum Vergleich und zur Evaluation von Subgruppen beschrieben um den Benutzer bei diesen essentiellen Schritten zu unterstützen. Weiterhin wird leicht zu formalisierendes Hintergrundwissen vorgestellt, das im Subgruppenentdeckungsprozess in vielfältiger Weise eingesetzt werden kann: Um den Entdeckungsprozess zu fokussieren, den Suchraum einzuschränken, und letztendlich die Effizienz der Entdeckungsmethode zu erhöhen. Insbesondere wird Hintergrundwissen eingeführt, um die Elemente der Anwendungsdomäne zu filtern, um geeignete Abstraktionen zu definieren, Werte zusammenzufassen, und die gefundenen Subgruppenmuster nachzubearbeiten. Schließlich werden diese Techniken in einen wissensintensiven Prozess integriert, der sowohl automatische als auch interaktive Methoden zur Subgruppenentdeckung einschließt. Die praktische Bedeutung des vorgestellten Ansatzes hängt stark von den verfügbaren Werkzeugen ab. Dazu wird das VIKAMINE System als hochintegrierte Umgebung für die wissensintensive aktive Subgruppenentdeckung präsentiert. Die Evaluation des Ansatzes besteht aus zwei Teilen: Hinsichtlich einer Evaluation von Effizienz und Effektivität der Verfahren wird eine experimentelle Evaluation mit synthetischen Daten vorgestellt. Für diesen Zweck wird ein neuartiger in der Arbeit entwickelter Datengenerator angewandt, der eine einfache und intuitive Spezifikation der Datencharakteristiken erlaubt. Für die Evaluation des Ansatzes wurden Daten erzeugt, die ähnliche Charakteristiken aufweisen wie die Daten des angestrebten Einsatzbereichs. Die Ergebnisse der Evaluation zeigen, dass der neuartige SD-Map Algorithmus den anderen in der Arbeit beschriebenen Standard-Algorithmen überlegen ist. Sowohl hinsichtlich der Effizienz, als auch von Precision/Recall bezogen auf die heuristischen Algorithmen bietet SD-Map deutliche Vorteile. Subjektive Evaluationskriterien sind durch die Benutzerakzeptanz, den Nutzen des Ansatzes, und die Interessantheit der Ergebnisse gegeben. Es werden fünf Fallstudien für den Einsatz der vorgestellten Techniken beschrieben: Der Ansatz wurde in medizinischen und technischen Anwendungen mit realen Daten eingesetzt. Dabei wurde er von den Benutzern sehr gut angenommen, und im praktischen Einsatz konnte neuartiges, nützliches, und interessantes Wissen entdeckt werden. KW - Data Mining KW - Algorithmus KW - Visualisierung KW - Subgruppenentdeckung KW - Hintergrundwissen KW - Wissensendeckung KW - Data Mining KW - Visualisierung KW - Subgroup Mining KW - Background Knowledge KW - Knowledge Discovery KW - Data Mining KW - Visualization Y1 - 2006 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-21004 ER - TY - JOUR A1 - Oberdörfer, Sebastian A1 - Latoschik, Marc Erich T1 - Knowledge encoding in game mechanics: transfer-oriented knowledge learning in desktop-3D and VR JF - International Journal of Computer Games Technology N2 - Affine Transformations (ATs) are a complex and abstract learning content. Encoding the AT knowledge in Game Mechanics (GMs) achieves a repetitive knowledge application and audiovisual demonstration. Playing a serious game providing these GMs leads to motivating and effective knowledge learning. Using immersive Virtual Reality (VR) has the potential to even further increase the serious game’s learning outcome and learning quality. This paper compares the effectiveness and efficiency of desktop-3D and VR in respect to the achieved learning outcome. Also, the present study analyzes the effectiveness of an enhanced audiovisual knowledge encoding and the provision of a debriefing system. The results validate the effectiveness of the knowledge encoding in GMs to achieve knowledge learning. The study also indicates that VR is beneficial for the overall learning quality and that an enhanced audiovisual encoding has only a limited effect on the learning outcome. KW - games Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-201159 VL - 2019 ER - TY - JOUR A1 - Fathy, Moustafa A1 - Darwish, Mostafa A. A1 - Abdelhamid, Al-Shaimaa M. A1 - Alrashedy, Gehad M. A1 - Othman, Othman Ali A1 - Naseem, Muhammad A1 - Dandekar, Thomas A1 - Othman, Eman M. T1 - Kinetin ameliorates cisplatin-induced hepatotoxicity and lymphotoxicity via attenuating oxidative damage, cell apoptosis and inflammation in rats JF - Biomedicines N2 - Though several previous studies reported the in vitro and in vivo antioxidant effect of kinetin (Kn), details on its action in cisplatin-induced toxicity are still scarce. In this study we evaluated, for the first time, the effects of kinetin in cisplatin (cp)- induced liver and lymphocyte toxicity in rats. Wistar male albino rats were divided into nine groups: (i) the control (C), (ii) groups 2,3 and 4, which received 0.25, 0.5 and 1 mg/kg kinetin for 10 days; (iii) the cisplatin (cp) group, which received a single intraperitoneal injection of CP (7.0 mg/kg); and (iv) groups 6, 7, 8 and 9, which received, for 10 days, 0.25, 0.5 and 1 mg/kg kinetin or 200 mg/kg vitamin C, respectively, and Cp on the fourth day. CP-injected rats showed a significant impairment in biochemical, oxidative stress and inflammatory parameters in hepatic tissue and lymphocytes. PCR showed a profound increase in caspase-3, and a significant decline in AKT gene expression. Intriguingly, Kn treatment restored the biochemical, redox status and inflammatory parameters. Hepatic AKT and caspase-3 expression as well as CD95 levels in lymphocytes were also restored. In conclusion, Kn mitigated oxidative imbalance, inflammation and apoptosis in CP-induced liver and lymphocyte toxicity; therefore, it can be considered as a promising therapy. KW - cisplatin KW - hepatotoxicity KW - lymphotoxicity KW - oxidative stress KW - AKT KW - CD95 KW - caspase-3 Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-281686 SN - 2227-9059 VL - 10 IS - 7 ER - TY - JOUR A1 - Kempf, Sebastian A1 - Krug, Markus A1 - Puppe, Frank T1 - KIETA: Key-insight extraction from scientific tables JF - Applied Intelligence N2 - An important but very time consuming part of the research process is literature review. An already large and nevertheless growing ground set of publications as well as a steadily increasing publication rate continue to worsen the situation. Consequently, automating this task as far as possible is desirable. Experimental results of systems are key-insights of high importance during literature review and usually represented in form of tables. Our pipeline KIETA exploits these tables to contribute to the endeavor of automation by extracting them and their contained knowledge from scientific publications. The pipeline is split into multiple steps to guarantee modularity as well as analyzability, and agnosticim regarding the specific scientific domain up until the knowledge extraction step, which is based upon an ontology. Additionally, a dataset of corresponding articles has been manually annotated with information regarding table and knowledge extraction. Experiments show promising results that signal the possibility of an automated system, while also indicating limits of extracting knowledge from tables without any context. KW - table extraction KW - table understanding KW - ontology KW - key-insight extraction KW - information extraction Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-324180 SN - 0924-669X VL - 53 IS - 8 ER - TY - THES A1 - Fehler, Manuel T1 - Kalibrierung Agenten-basierter Simulationen T1 - Calibration of Agent-based Simulations N2 - In der vorliegenden Arbeit wird das Problem der Kalibrierung Agenten-basierter Simulationen (ABS) behandelt, also das Problem, die Parameterwerte eines Agenten-basierten Simulationsmodells so einzustellen, dass valides Simulationsverhalten erreicht wird. Das Kalibrierungsproblem für Simulationen an sich ist nicht neu und ist im Rahmen klassischer Simulationsparadigmen, wie z.B. der Makro-Simulation, fester Bestandteil der Forschung. Im Vergleich zu den dort betrachteten Kalibrierungsproblemen zeichnet sich das Kalibrierungsproblem für ABS jedoch durch eine Reihe zusätzlicher Herausforderungen aus, welche die direkte Anwendung existierender Kalibrierungsverfahren in begrenzter Zeit erschweren, bzw. nicht mehr sinnvoll zulassen. Die Lösung dieser Probleme steht im Zentrum dieser Dissertation: Das Ziel besteht darin, den Nutzer bei der Kalibrierung von ABS auf der Basis von unzureichenden, potentiell fehlerhaften Daten und Wissen zu unterstützen. Dabei sollen drei Hauptprobleme gelöst werden: 1)Vereinfachung der Kalibrierung großer Agenten-Parametermengen auf der Mikro- Ebene in Agenten-basierten Simulationen durch Ausnutzung der spezifischen Struktur von ABS (nämlich dem Aufbau aus einer Menge von Agentenmodellen). 2)Kalibrierung Agenten-basierter Simulationen, so dass auf allen relevanten Beobachtungsebenen valides Simulationsverhalten erzeugt wird (mindestens Mikro und Makro-Ebene). Als erschwerende Randbedingung muss die Kalibrierung unter der Voraussetzung einer Makro-Mikro-Wissenslücke durchgeführt werden. 3)Kalibrierung Agenten-basierter Simulationen auf der Mikro-Ebene unter der Voraussetzung, dass zur Kalibrierung einzelner Agentenmodelle nicht ausreichend und potentiell verfälschte Daten zur Verhaltensvalidierung zur Verfügung stehen. Hierzu wird in dieser Arbeit das sogenannte Makro-Mikro-Verfahren zur Kalibrierung von Agenten-basierten Simulationen entwickelt. Das Verfahren besteht aus einem Basisverfahren, das im Verlauf der Arbeit um verschiedene Zusatzverfahren erweitert wird. Das Makro-Mikro-Verfahren und seine Erweiterungen sollen dazu dienen, die Modellkalibrierung trotz stark verrauschter Daten und eingeschränktem Wissen über die Wirkungszusammenhänge im Originalsystem geeignet zu ermöglichen und dabei den Kalibrierungsprozess zu beschleunigen: 1) Makro-Mikro-Kalibrierungsverfahren: Das in dieser Arbeit entwickelte Makro- Mikro-Verfahren unterstützt den Nutzer durch eine kombinierte Kalibrierung auf der Mikro- und der Makro-Beobachtungsebene, die gegebenenfalls durch Zwischenebenen erweitert werden kann. Der Grundgedanke des Verfahrens besteht darin, das Kalibrierungsproblem in eines auf aggregierter Verhaltensebene und eines auf der Ebene des Mikro-Agentenverhaltens aufzuteilen. Auf der Makro-Ebene wird nach validen idealen aggregierten Verhaltensmodellen (IVM) der Agenten gesucht. Auf der Mikro-Ebene wird versucht die individuellen Modelle der Agenten auf Basis des erwünschten Gesamtverhaltens und der ermittelten IVM so zu kalibrieren, das insgesamt Simulationsverhalten entsteht, das sowohl auf Mikro- als auch auf Makro-Ebene valide ist. 2) Erweiterung 1: Robuste Kalibrierung: Um den Umgang mit potentiell verrauschten Validierungskriterien (d.h. mit verrauschten Daten über ein Originalsystem, auf denen die Validierungskriterien der Simulation beruhen) und Modellteilen während der Kalibrierung von ABS zu ermöglichen, wird eine robuste Kalibrierungstechnik zur Anwendung im Makro-Mikro-Verfahren entwickelt. 3) Erweiterung 2: Kalibrierung mit Heterogenitätssuche: Als zweite Erweiterung des Makro-Mikro-Verfahrens wird ein Verfahren entwickelt, das das Problem des unklaren Detaillierungsgrades von ABS auf der Ebene der Parameterwerte adressiert. Prinzipiell kann zwar jeder Agent unterschiedliche Parameterwerte verwenden, obwohl eine geringere Heterogenität zur Erzeugung validen Verhaltens ausreichend wäre. Die entwickelte Erweiterung versucht, während der Kalibrierung, eine geeignete Heterogenitätsausprägung für die Parameterwerte der Agenten zu ermitteln. Unter einer Heterogenitätsausprägung wird dabei eine Einteilung der simulierten Agenten in Gruppen mit jeweils gleichen Parameterwerten verstanden. Die Heterogenitätssuche dient dazu, einen Kompromiss zu finden zwischen der Notwendigkeit, sehr große Parametersuchräume durchsuchen zu müssen und gleichzeitig den Suchraum so klein wie möglich halten zu wollen. N2 - In this doctoral thesis the problem of calibrating agent-based simulations (ABS) is treated, i.e. the problem to adjust the parameter values of an agent-based simulation model to achieve valid simulation behavior. The calibration problem for simulations per se is not new and is an active part of research in the context of traditional simulation paradigms, such as the macro-simulation. Compared to the problems considered there the problems for ABS can be distinguished by several additional challenges that complicate the direct application of existing calibration procedures in a limited time, or challenges that do not allow applying existing procedures at all. The goal of this thesis is to assist the user in the calibration of ABS on the basis of incomplete and potentially noisy data or knowledge and in dealing with large amounts of parameter values if an ABS with many individual agents needs to be calibrated. The thesis covers the following three main topics: 1) Simplification of the calibration of many agent parameter values on the micro-level in ABS. This is done by exploiting the specific structure of ABS (i.e. that an ABS constitutes of a lattice of agent models). 2) Calibration of agent-based simulations, so that valid simulation behavior is created on all relevant behavior observation levels (at least micro- and macro-level). This needs to be possible without having full knowledge about how the macro observation level behavior constitutes from the modeled micro behavior. 3) Calibration of agent-based simulations on the micro-level under the constraint that only partial and potentially noisy data for testing and validation of single individual agent models is available. To achieve this the so-called “Macro-Micro Procedure” for calibrating agent-based simulations is developed. The approach consists of a basic procedure that is extended in the course of the work with various additional techniques: 1)Macro-Micro-Calibration Procedure: The Macro-Micro Procedure supports the user by applying a combined calibration on the micro and the macro-observation level, which can optionally be expanded using additional intermediate levels. The basic idea of the procedure consists of separating the calibration problem into one at the aggregate behavior level and one at the level of the micro-agent behavior. At the macro level, valid simulation behavior for ideal aggregate behavior models (IAM) of agents is being determined. At the micro level, the goal is to calibrate the models of the individual agents based on the desired overall behavior and the determined IAM from the macro level. Upon completion the simulation behavior created shall be valid both at the micro and also at a macro level. 2)Extension 1: Robust Calibration: In order to deal with potentially noisy validation criteria and model parts (i.e. with noisy data about the original system from which the validation criteria of the simulation are created) a robust calibration technique is developed that can be used as part of the Macro-Micro-Procedure. 3)Extension 2: Calibration with heterogeneity search: The second extension of the Macro-Micro-Procedure addresses the problem of an unclear level of detail on the level of the parameter values. Theoretically it is possible to use different parameter values for each individual simulated agent which leads to a huge parameter search space. Often it is however sufficient to use a lower heterogeneity in the parameter values to generate valid behavior which would allow calibration in a smaller search space. The developed extension attempts to determine such a suitable heterogeneity manifestation for the parameter values of the agents as part of the calibration process itself. A heterogeneity manifestation is performed by dividing the agents into groups of agents with homogenous parameter values. The developed heterogeneity search offers a compromise between the necessity of having to search very large parameter search spaces and the goal to keep the search space as small as possible during the calibration. KW - Computersimulation KW - Mehrebenensimulation KW - Autonomer Agent KW - Agenten-basierte Simulation KW - Multiagentensimulation KW - Parameterkalibrierung KW - Hierarchische Simulation KW - Simulation KW - Agent KW - Calibration KW - Optimization KW - Agent-based Simulation KW - Multi-Agent-Simulation Y1 - 2010 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-64762 ER - TY - JOUR A1 - Steinhaeusser, Sophia C. A1 - Oberdörfer, Sebastian A1 - von Mammen, Sebastian A1 - Latoschik, Marc Erich A1 - Lugrin, Birgit T1 - Joyful adventures and frightening places – designing emotion-inducing virtual environments JF - Frontiers in Virtual Reality N2 - Virtual environments (VEs) can evoke and support emotions, as experienced when playing emotionally arousing games. We theoretically approach the design of fear and joy evoking VEs based on a literature review of empirical studies on virtual and real environments as well as video games’ reviews and content analyses. We define the design space and identify central design elements that evoke specific positive and negative emotions. Based on that, we derive and present guidelines for emotion-inducing VE design with respect to design themes, colors and textures, and lighting configurations. To validate our guidelines in two user studies, we 1) expose participants to 360° videos of VEs designed following the individual guidelines and 2) immerse them in a neutral, positive and negative emotion-inducing VEs combining all respective guidelines in Virtual Reality. The results support our theoretically derived guidelines by revealing significant differences in terms of fear and joy induction. KW - virtual reality KW - virtual environments KW - immersion KW - emotions KW - design Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-284831 SN - 2673-4192 VL - 3 ER - TY - JOUR A1 - Landeck, Maximilian A1 - Alvarez Igarzábal, Federico A1 - Unruh, Fabian A1 - Habenicht, Hannah A1 - Khoshnoud, Shiva A1 - Wittmann, Marc A1 - Lugrin, Jean-Luc A1 - Latoschik, Marc Erich T1 - Journey through a virtual tunnel: Simulated motion and its effects on the experience of time JF - Frontiers in Virtual Reality N2 - This paper examines the relationship between time and motion perception in virtual environments. Previous work has shown that the perception of motion can affect the perception of time. We developed a virtual environment that simulates motion in a tunnel and measured its effects on the estimation of the duration of time, the speed at which perceived time passes, and the illusion of self-motion, also known as vection. When large areas of the visual field move in the same direction, vection can occur; observers often perceive this as self-motion rather than motion of the environment. To generate different levels of vection and investigate its effects on time perception, we developed an abstract procedural tunnel generator. The generator can simulate different speeds and densities of tunnel sections (visibly distinguishable sections that form the virtual tunnel), as well as the degree of embodiment of the user avatar (with or without virtual hands). We exposed participants to various tunnel simulations with different durations, speeds, and densities in a remote desktop and a virtual reality (VR) laboratory study. Time passed subjectively faster under high-speed and high-density conditions in both studies. The experience of self-motion was also stronger under high-speed and high-density conditions. Both studies revealed a significant correlation between the perceived passage of time and perceived self-motion. Subjects in the virtual reality study reported a stronger self-motion experience, a faster perceived passage of time, and shorter time estimates than subjects in the desktop study. Our results suggest that a virtual tunnel simulation can manipulate time perception in virtual reality. We will explore these results for the development of virtual reality applications for therapeutic approaches in our future work. This could be particularly useful in treating disorders like depression, autism, and schizophrenia, which are known to be associated with distortions in time perception. For example, the tunnel could be therapeutically applied by resetting patients’ time perceptions by exposing them to the tunnel under different conditions, such as increasing or decreasing perceived time. KW - passage of time KW - illusion of self-motion KW - vection KW - virtual tunnel KW - therapeutic application KW - virtual reality KW - extended reality (XR) Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-301519 SN - 2673-4192 VL - 3 ER - TY - JOUR A1 - Karl, Stefan A1 - Dandekar, Thomas T1 - Jimena: Efficient computing and system state identification for genetic regulatory networks JF - BMC Bioinformatics N2 - Background: Boolean networks capture switching behavior of many naturally occurring regulatory networks. For semi-quantitative modeling, interpolation between ON and OFF states is necessary. The high degree polynomial interpolation of Boolean genetic regulatory networks (GRNs) in cellular processes such as apoptosis or proliferation allows for the modeling of a wider range of node interactions than continuous activator-inhibitor models, but suffers from scaling problems for networks which contain nodes with more than ~10 inputs. Many GRNs from literature or new gene expression experiments exceed those limitations and a new approach was developed. Results: (i) As a part of our new GRN simulation framework Jimena we introduce and setup Boolean-tree-based data structures; (ii) corresponding algorithms greatly expedite the calculation of the polynomial interpolation in almost all cases, thereby expanding the range of networks which can be simulated by this model in reasonable time. (iii) Stable states for discrete models are efficiently counted and identified using binary decision diagrams. As application example, we show how system states can now be sampled efficiently in small up to large scale hormone disease networks (Arabidopsis thaliana development and immunity, pathogen Pseudomonas syringae and modulation by cytokinins and plant hormones). Conclusions: Jimena simulates currently available GRNs about 10-100 times faster than the previous implementation of the polynomial interpolation model and even greater gains are achieved for large scale-free networks. This speed-up also facilitates a much more thorough sampling of continuous state spaces which may lead to the identification of new stable states. Mutants of large networks can be constructed and analyzed very quickly enabling new insights into network robustness and behavior. KW - Boolean function KW - genetic regulatory network KW - interpolation KW - stable state KW - binary decision diagram KW - Boolean tree Y1 - 2013 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-128671 VL - 14 ER - TY - RPRT A1 - Rauber, Christof A. O. A1 - Brechtel, Lukas A1 - Schotten, Hans D. T1 - JCAS-Enabled Sensing as a Service in 6th-Generation Mobile Communication Networks T2 - KuVS Fachgespräch - Würzburg Workshop on Modeling, Analysis and Simulation of Next-Generation Communication Networks 2023 (WueWoWAS’23) N2 - The introduction of new types of frequency spectrum in 6G technology facilitates the convergence of conventional mobile communications and radar functions. Thus, the mobile network itself becomes a versatile sensor system. This enables mobile network operators to offer a sensing service in addition to conventional data and telephony services. The potential benefits are expected to accrue to various stakeholders, including individuals, the environment, and society in general. The paper discusses technological development, possible integration, and use cases, as well as future development areas. KW - Sensing-aaS KW - JCAS KW - 6G Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-322135 ER - TY - INPR A1 - Nassourou, Mohamadou T1 - Java Web Frameworks Which One to Choose? N2 - This article discusses web frameworks that are available to a software developer in Java language. It introduces MVC paradigm and some frameworks that implement it. The article presents an overview of Struts, Spring MVC, JSF Frameworks, as well as guidelines for selecting one of them as development environment. KW - Java Frameworks KW - MVC KW - Struts KW - Spring KW - JSF Y1 - 2010 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-49407 ER - TY - GEN A1 - Funken, Matthias A1 - Tscherner, Michael T1 - Jahresbericht 2018 des Rechenzentrums der Universität Würzburg T1 - Annual Report 2018 of the Computer Center, University of Wuerzburg N2 - Eine Übersicht über die Aktivitäten des Rechenzentrums im Jahr 2018. T3 - Jahresbericht des Rechenzentrums der Universität Würzburg - 2018 KW - Julius-Maximilians-Universität Würzburg KW - Jahresbericht KW - Jahresbericht KW - Rechenzentrum KW - RZUW KW - annual report KW - Computer Center University of Wuerzburg Y1 - 2019 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-188265 UR - https://www.rz.uni-wuerzburg.de/wir/publikationen/ ET - 1. Auflage ER - TY - GEN A1 - Funken, Matthias A1 - Tscherner, Michael T1 - Jahresbericht 2017 des Rechenzentrums der Universität Würzburg T1 - Annual Report 2017 of the Computer Center, University of Wuerzburg N2 - Eine Übersicht über die Aktivitäten des Rechenzentrums im Jahr 2017. T3 - Jahresbericht des Rechenzentrums der Universität Würzburg - 2017 KW - Julius-Maximilians-Universität Würzburg KW - RZUW KW - Jahresbericht KW - Rechenzentrum KW - Computer Center University of Wuerzburg KW - annual report Y1 - 2018 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-168537 UR - https://www.rz.uni-wuerzburg.de/wir/publikationen/ ET - 1. Auflage ER - TY - GEN T1 - Jahresbericht 2016 des Rechenzentrums der Universität Würzburg T1 - Annual Report 2016 of the Computer Center, University of Wuerzburg N2 - Das Dokument umfasst eine jährliche Zusammenfassung der Aktivitäten des Rechenzentrums als zentraler IT-Dienstleister der Universität Würzburg T3 - Jahresbericht des Rechenzentrums der Universität Würzburg - 2016 KW - Jahresbericht KW - Julius-Maximilians-Universität Würzburg KW - Rechenzentrum KW - annual report KW - Computer Center University of Wuerzburg KW - RZUW Y1 - 2017 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-153558 UR - https://www.rz.uni-wuerzburg.de/wir/publikationen/ ET - 1. Auflage ER - TY - GEN T1 - Jahresbericht 2015 T1 - Annual Report 2015 N2 - Jahresbericht 2015 des Rechenzentrums der Universität Würzburg N2 - Annual Report 2015 of the Computer Center, University of Wuerzburg T3 - Jahresbericht des Rechenzentrums der Universität Würzburg - 2015 KW - Julius-Maximilians-Universität Würzburg. Rechenzentrum KW - annual report KW - Computer Center University of Wuerzburg KW - Jahresbericht KW - RZUW Y1 - 2016 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-136599 UR - https://www.rz.uni-wuerzburg.de/wir/publikationen/ ER - TY - GEN T1 - Jahresbericht 2014 T1 - Annual Report 2014 N2 - Jahresbericht 2014 des Rechenzentrums der Universität Würzburg N2 - Annual Report 2014 of the Computer Center, University of Wuerzburg T3 - Jahresbericht des Rechenzentrums der Universität Würzburg - 2014 KW - Rechenzentrum Universität Würzburg KW - annual report KW - Computer Center University of Wuerzburg KW - Jahresbericht Y1 - 2016 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-124432 UR - https://www.rz.uni-wuerzburg.de/infos/publikationen/ ER - TY - JOUR A1 - Krupitzer, Christian A1 - Eberhardinger, Benedikt A1 - Gerostathopoulos, Ilias A1 - Raibulet, Claudia T1 - Introduction to the special issue “Applications in Self-Aware Computing Systems and their Evaluation” JF - Computers N2 - The joint 1st Workshop on Evaluations and Measurements in Self-Aware Computing Systems (EMSAC 2019) and Workshop on Self-Aware Computing (SeAC) was held as part of the FAS* conference alliance in conjunction with the 16th IEEE International Conference on Autonomic Computing (ICAC) and the 13th IEEE International Conference on Self-Adaptive and Self-Organizing Systems (SASO) in Umeå, Sweden on 20 June 2019. The goal of this one-day workshop was to bring together researchers and practitioners from academic environments and from the industry to share their solutions, ideas, visions, and doubts in self-aware computing systems in general and in the evaluation and measurements of such systems in particular. The workshop aimed to enable discussions, partnerships, and collaborations among the participants. This special issue follows the theme of the workshop. It contains extended versions of workshop presentations as well as additional contributions. KW - self-aware computing systems KW - quality evaluation KW - measurements KW - quality assurance KW - autonomous KW - self-adaptive KW - self-managing systems Y1 - 2020 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-203439 SN - 2073-431X VL - 9 IS - 1 ER -