TY - JOUR A1 - Marquardt, André A1 - Hartrampf, Philipp A1 - Kollmannsberger, Philip A1 - Solimando, Antonio G. A1 - Meierjohann, Svenja A1 - Kübler, Hubert A1 - Bargou, Ralf A1 - Schilling, Bastian A1 - Serfling, Sebastian E. A1 - Buck, Andreas A1 - Werner, Rudolf A. A1 - Lapa, Constantin A1 - Krebs, Markus T1 - Predicting microenvironment in CXCR4- and FAP-positive solid tumors — a pan-cancer machine learning workflow for theranostic target structures JF - Cancers N2 - (1) Background: C-X-C Motif Chemokine Receptor 4 (CXCR4) and Fibroblast Activation Protein Alpha (FAP) are promising theranostic targets. However, it is unclear whether CXCR4 and FAP positivity mark distinct microenvironments, especially in solid tumors. (2) Methods: Using Random Forest (RF) analysis, we searched for entity-independent mRNA and microRNA signatures related to CXCR4 and FAP overexpression in our pan-cancer cohort from The Cancer Genome Atlas (TCGA) database — representing n = 9242 specimens from 29 tumor entities. CXCR4- and FAP-positive samples were assessed via StringDB cluster analysis, EnrichR, Metascape, and Gene Set Enrichment Analysis (GSEA). Findings were validated via correlation analyses in n = 1541 tumor samples. TIMER2.0 analyzed the association of CXCR4 / FAP expression and infiltration levels of immune-related cells. (3) Results: We identified entity-independent CXCR4 and FAP gene signatures representative for the majority of solid cancers. While CXCR4 positivity marked an immune-related microenvironment, FAP overexpression highlighted an angiogenesis-associated niche. TIMER2.0 analysis confirmed characteristic infiltration levels of CD8+ cells for CXCR4-positive tumors and endothelial cells for FAP-positive tumors. (4) Conclusions: CXCR4- and FAP-directed PET imaging could provide a non-invasive decision aid for entity-agnostic treatment of microenvironment in solid malignancies. Moreover, this machine learning workflow can easily be transferred towards other theranostic targets. KW - machine learning KW - tumor microenvironment KW - immune infiltration KW - angiogenesis KW - mRNA KW - miRNA KW - transcriptome Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-305036 SN - 2072-6694 VL - 15 IS - 2 ER - TY - JOUR A1 - Kazuhino, Koshino A1 - Werner, Rudolf A. A1 - Toriumi, Fuijo A1 - Javadi, Mehrbod S. A1 - Pomper, Martin G. A1 - Solnes, Lilja B. A1 - Verde, Franco A1 - Higuchi, Takahiro A1 - Rowe, Steven P. T1 - Generative Adversarial Networks for the Creation of Realistic Artificial Brain Magnetic Resonance Images JF - Tomography N2 - Even as medical data sets become more publicly accessible, most are restricted to specific medical conditions. Thus, data collection for machine learning approaches remains challenging, and synthetic data augmentation, such as generative adversarial networks (GAN), may overcome this hurdle. In the present quality control study, deep convolutional GAN (DCGAN)-based human brain magnetic resonance (MR) images were validated by blinded radiologists. In total, 96 T1-weighted brain images from 30 healthy individuals and 33 patients with cerebrovascular accident were included. A training data set was generated from the T1-weighted images and DCGAN was applied to generate additional artificial brain images. The likelihood that images were DCGAN-created versus acquired was evaluated by 5 radiologists (2 neuroradiologists [NRs], vs 3 non-neuroradiologists [NNRs]) in a binary fashion to identify real vs created images. Images were selected randomly from the data set (variation of created images, 40%-60%). None of the investigated images was rated as unknown. Of the created images, the NRs rated 45% and 71% as real magnetic resonance imaging images (NNRs, 24%, 40%, and 44%). In contradistinction, 44% and 70% of the real images were rated as generated images by NRs (NNRs, 10%, 17%, and 27%). The accuracy for the NRs was 0.55 and 0.30 (NNRs, 0.83, 0.72, and 0.64). DCGAN-created brain MR images are similar enough to acquired MR images so as to be indistinguishable in some cases. Such an artificial intelligence algorithm may contribute to synthetic data augmentation for "data-hungry" technologies, such as supervised machine learning approaches, in various clinical applications. KW - AI KW - Magnetresonanztomografie KW - artificial intelligence KW - magnetic resonance imaging KW - MRI KW - DCGAN KW - GAN KW - stroke KW - machine learning Y1 - 2018 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-172185 VL - 4 IS - 4 ER - TY - JOUR A1 - Werner, Rudolf A. A1 - Higuchi, Takahiro A1 - Nose, Naoko A1 - Toriumi, Fujio A1 - Matsusaka, Yohji A1 - Kuji, Ichiei A1 - Kazuhiro, Koshino T1 - Generative adversarial network-created brain SPECTs of cerebral ischemia are indistinguishable to scans from real patients JF - Scientific reports N2 - Deep convolutional generative adversarial networks (GAN) allow for creating images from existing databases. We applied a modified light-weight GAN (FastGAN) algorithm to cerebral blood flow SPECTs and aimed to evaluate whether this technology can generate created images close to real patients. Investigating three anatomical levels (cerebellum, CER; basal ganglia, BG; cortex, COR), 551 normal (248 CER, 174 BG, 129 COR) and 387 pathological brain SPECTs using N-isopropyl p-I-123-iodoamphetamine (123I-IMP) were included. For the latter scans, cerebral ischemic disease comprised 291 uni- (66 CER, 116 BG, 109 COR) and 96 bilateral defect patterns (44 BG, 52 COR). Our model was trained using a three-compartment anatomical input (dataset ‘A’; including CER, BG, and COR), while for dataset ‘B’, only one anatomical region (COR) was included. Quantitative analyses provided mean counts (MC) and left/right (LR) hemisphere ratios, which were then compared to quantification from real images. For MC, ‘B’ was significantly different for normal and bilateral defect patterns (P < 0.0001, respectively), but not for unilateral ischemia (P = 0.77). Comparable results were recorded for LR, as normal and ischemia scans were significantly different relative to images acquired from real patients (P ≤ 0.01, respectively). Images provided by ‘A’, however, revealed comparable quantitative results when compared to real images, including normal (P = 0.8) and pathological scans (unilateral, P = 0.99; bilateral, P = 0.68) for MC. For LR, only uni- (P = 0.03), but not normal or bilateral defect scans (P ≥ 0.08) reached significance relative to images of real patients. With a minimum of only three anatomical compartments serving as stimuli, created cerebral SPECTs are indistinguishable to images from real patients. The applied FastGAN algorithm may allow to provide sufficient scan numbers in various clinical scenarios, e.g., for “data-hungry” deep learning technologies or in the context of orphan diseases. KW - computational biology and bioinformatics KW - machine learning KW - neurology Y1 - 2022 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:bvb:20-opus-300757 VL - 12 ER -