Refine
Has Fulltext
- yes (73)
Is part of the Bibliography
- yes (73)
Year of publication
Document Type
- Journal article (73) (remove)
Language
- English (73) (remove)
Keywords
- Gaussian approximation (3)
- Mathematik (3)
- count time series (3)
- nonsmooth optimization (3)
- optimal control (3)
- Anaphylaxis (2)
- BGK approximation (2)
- Fokker–Planck equation (2)
- Grundvorstellung (2)
- HIV (2)
- basic mental models (2)
- cardinality constraints (2)
- estimation error (2)
- kinetic model (2)
- mathematical modelling (2)
- mathematics (2)
- multi-fluid mixture (2)
- numerical analysis (2)
- optimal control theory (2)
- proximal gradient method (2)
- trabeculectomy (2)
- viral load (2)
- (approximate) functional equation (1)
- 5-fluorouracil (1)
- AIDS (1)
- Ableitung (1)
- Bayesian approach (1)
- Bayesian inverse problems (1)
- Bernstein-type inequality (1)
- Bi-Fidelity method (1)
- Black Scholes equation (1)
- Bloom setting (1)
- Brockett (1)
- CSF (1)
- Caputo fractional derivative (1)
- Carleson embedding theorem (1)
- Cauchy-Born rule (1)
- Cayley graph (1)
- D-norm (1)
- DAT (1)
- DNA replication (1)
- Discrete-to-continuum limits (1)
- Double sensitization (1)
- Drug allergy (1)
- Drug reaction (1)
- EJMA-D-19-00287 (1)
- Euler system (1)
- Euler–Bernoulli damped beam (1)
- Field sting (1)
- Gamma-convergence (1)
- Gram points (1)
- Gram’s law (1)
- Graph eigenvalues (1)
- Graph products (1)
- Graphnullity (1)
- HAND (1)
- HIV infections (1)
- Hautus test (1)
- Hecke L-functions (1)
- Hecke eigenforms (1)
- Honey bee (1)
- Hurwitz zeta function (1)
- Hymenoptera venom (1)
- Immunotherapy (1)
- Julia line (1)
- Keller–Segel model (1)
- Koenigs function (1)
- Krylow matrix (1)
- Laplace approximation (1)
- Lerch zeta function (1)
- Lie groups (1)
- Lie n-algebroids (1)
- Local rings (1)
- Lotka-Volterra models (1)
- Lyapunov functions (1)
- MIND estimator (1)
- MSC 11M35 (1)
- Martensit (1)
- Mastocytosis (1)
- Mehrskalenmodell (1)
- Nash bargaining problem (1)
- Nash equilibria (1)
- Nash equilibrium (1)
- Newton methods (1)
- Non-steroidal anti-inflammatory drug (1)
- Phasenumwandlung (1)
- Poisson algebras (1)
- Pontryagin maximum principle (1)
- Pseudo-allergy (1)
- Razumikhin method (1)
- Reachability matrix (1)
- Relapse (1)
- Riemann hypothesis (1)
- Riemann zeta-function (1)
- Risk factor (1)
- STEM classroom (1)
- STEM education (1)
- STEM integration (1)
- Struktur (1)
- T3s (1)
- Testinstrument (1)
- Treatment failure (1)
- Value at Risk (1)
- Vandermonde matrix (1)
- Variationsrechnung (1)
- Vespula (1)
- Von Mises conditions (1)
- Weibull type density (1)
- Zero-divisor graphs (1)
- adaptive intervention competence (1)
- adjoint and coadjoint representations (1)
- algebraic degree (1)
- approaches in textbooks (1)
- asthma (1)
- asymptotic analysis (1)
- asymptotic sufficiency (1)
- asymptotically flat ends (1)
- atomistic models (1)
- augmented Lagrange method (1)
- augmented Lagrangian (1)
- augmented Lagrangian method (1)
- avoidance (1)
- backward orbit (1)
- basic mental model (1)
- bevacizumab (1)
- bicommutators (1)
- bilinear evolution model (1)
- bleb failure (1)
- bleb scarring (1)
- blood (1)
- body weight (1)
- bounded input bounded output stability (1)
- canal surgery (1)
- canaloplasty (1)
- cancer (1)
- cataract surgery (1)
- chlamydia (1)
- chlamydia infection (1)
- chlamydia trachomatis (1)
- circumferential viscodilation (1)
- co-extremality coefficient (1)
- coherent forecasting (1)
- companion matrix (1)
- complex polynomials (1)
- composition of functions (1)
- compressible fluid (1)
- computational biology and bioinformatics (1)
- condenser capacity (1)
- conecontinuity type constraint qualification (1)
- conservation laws (1)
- continuum limit (1)
- copula (1)
- critical line (1)
- cubic-monoclinic martensites (1)
- cyclic matrix (1)
- degrees of freedom in internal energy (1)
- derivative (1)
- diagnosis (1)
- differential games (1)
- differential graded modules (1)
- differential nash games (1)
- diffuse interface models (1)
- digital technologies (1)
- digital tools (1)
- discrete systems (1)
- dopamine (1)
- drug adherence (1)
- dyadic product BMO (1)
- echocardiography (1)
- elliptic PDE (1)
- elliptic problems (1)
- empirical evidence (1)
- empirical investigation (1)
- ensemble optimal control problems (1)
- ensemble reachability (1)
- entropy minimization (1)
- examples and counterexamples (1)
- exchangeable D-norms (1)
- existence of solutions (1)
- expected shortfall (1)
- expectiles (1)
- exponent pairs (1)
- exponential growth (1)
- extremal exchangeable D-norms (1)
- extreme order statistics (1)
- extreme value distribution (1)
- extreme value theory (1)
- finite differences (1)
- finite groups (1)
- flexible microcatheter (1)
- fluid–structure interaction model (1)
- foliation (1)
- fredholm operator (1)
- functional thinking (1)
- generalized Pareto distribution (1)
- generator of D-norm (1)
- geometrically linear elasticity (1)
- geometrisch lineare Elastizitätstheorie (1)
- glaucoma (1)
- glaucoma surgery (1)
- global attractor (1)
- global convergence (1)
- green energy (1)
- growth models (1)
- harmonic measure (1)
- highly-active antiretroviral therapy (1)
- homogenization (1)
- homogenous parameter-depedent systems (1)
- hyperbolic area (1)
- hyperbolic systems (1)
- image denoising (1)
- immune activation (1)
- impulsive systems (1)
- infinite-dimensional systems (1)
- input-to-state stability (1)
- integral (1)
- integral graph (1)
- intermediate order statistics (1)
- inverse problems (1)
- jump-diffusion processes (1)
- kinetic chemotaxis equation (1)
- kubisch-monokliner Phasenübergang (1)
- large-time behaviour (1)
- literature review (1)
- local asymptotic normality (1)
- local classes (1)
- local existence (1)
- local input-to-state stability (1)
- lonlinear reaction-diffusion equations (1)
- lyapunov methods (1)
- macrophages (1)
- magnetic fluids (1)
- mathematical biology (1)
- mathematics classrooms (1)
- mid quantiles (1)
- mitomycin C (1)
- months follow-up (1)
- morbid obesity (1)
- multigrid schemes (1)
- multiscale methods (1)
- multiscale modeling (1)
- multivariate Gaussian distribution (1)
- multivariate exceedance (1)
- multivariate extreme value theory (1)
- multivariate max-domain of attraction (1)
- multivariate statistical process control (SPC) (1)
- mutually permutable (1)
- non-Lipschitz optimization (1)
- non-smooth and non-convex optimization (1)
- non-smooth large-scale optimisation (1)
- non-smooth optimal control problems (1)
- nonderogatory matrix (1)
- nonlinear inverse problems (1)
- nonlinear systems (1)
- numerical finance (1)
- obesity (1)
- open-angle glaucoma (1)
- operative principle (1)
- optimal control problem (1)
- order of growth (1)
- over-determined problem (1)
- p-soluble groups (1)
- p-supersolubility (1)
- partial integro-differential Fokker-Planck Equation (1)
- pedagogical content knowledge (1)
- pedestrian motion (1)
- petal (1)
- phacocanaloplasty (1)
- phacotrabeculectomy (1)
- phase I (1)
- phase II (1)
- plasma physics (1)
- polymerase chain reaction (1)
- polymorphism (1)
- polynomial chaos (1)
- pontryagin maximum principle (1)
- prediction interval (1)
- predictive performance (1)
- proximal method (1)
- proximal methods (1)
- pulmonary function (1)
- pulmonary hypertension (1)
- quantile forecasts (1)
- quantum evolution models (1)
- quasinormality constraint qualification (1)
- relaxation method (1)
- representations up to homotopy (1)
- robustness (1)
- role of mathematics in STEM (1)
- secondary education (1)
- semi-convex hulls (1)
- semi-konvexe Hüllen (1)
- semigroup of holomorphic functions (1)
- semilinear elliptic operators (1)
- semismooth Newton method (1)
- sequential optimality condition (1)
- sequential quadratic hamiltonian method (1)
- series (1)
- simulation (1)
- sparse control problems (1)
- stability (1)
- star products (1)
- state constraints (1)
- stationarity (1)
- stochastic processes (1)
- structure (1)
- successive approximations strategy (1)
- surgical outcomes (1)
- systems biology (1)
- tail conditional expectation (1)
- task design (1)
- teacher education (1)
- technology (1)
- test instrument (1)
- therapy (1)
- time‐varying delay (1)
- uncertain volatility (1)
- uncertainty quantification (1)
- value at risk (1)
- value-distribution (1)
- variational estimation (1)
- variational fracture (1)
- velocity jump process (1)
- velocity-dependent collision frequency (1)
- viral replication (1)
- weight loss (1)
- zeta-functions (1)
Institute
- Institut für Mathematik (73) (remove)
EU-Project number / Contract (GA) number
- 304617 (2)
This paper is devoted to a theoretical and numerical investigation of Nash equilibria and Nash bargaining problems governed by bilinear (input-affine) differential models. These systems with a bilinear state-control structure arise in many applications in, e.g., biology, economics, physics, where competition between different species, agents, and forces needs to be modelled. For this purpose, the concept of Nash equilibria (NE) appears appropriate, and the building blocks of the resulting differential Nash games are different control functions associated with different players that pursue different non-cooperative objectives. In this framework, existence of Nash equilibria is proved and computed with a semi-smooth Newton scheme combined with a relaxation method. Further, a related Nash bargaining (NB) problem is discussed. This aims at determining an improvement of all players’ objectives with respect to the Nash equilibria. Results of numerical experiments successfully demonstrate the effectiveness of the proposed NE and NB computational framework.
Providing adaptive, independence-preserving and theory-guided support to students in dealing with real-world problems in mathematics lessons is a major challenge for teachers in their professional practice. This paper examines this challenge in the context of simulations and mathematical modelling with digital tools: in addition to mathematical difficulties when autonomously working out individual solutions, students may also experience challenges when using digital tools. These challenges need to be closely examined and diagnosed, and might – if necessary – have to be overcome by intervention in such a way that the students can subsequently continue working independently. Thus, if a difficulty arises in the working process, two knowledge dimensions are necessary in order to provide adapted support to students. For teaching simulations and mathematical modelling with digital tools, more specifically, these knowledge dimensions are: pedagogical content knowledge about simulation and modelling processes supported by digital tools (this includes knowledge about phases and difficulties in the working process) and pedagogical content knowledge about interventions during the mentioned processes (focussing on characteristics of suitable interventions as well as their implementation and effects on the students’ working process). The two knowledge dimensions represent cognitive dispositions as the basis for the conceptualisation and operationalisation of a so-called adaptive intervention competence for teaching simulations and mathematical modelling with digital tools. In our article, we present a domain-specific process model and distinguish different types of teacher interventions. Then we describe the design and content of a university course at two German universities aiming to promote this domain-specific professional adaptive intervention competence, among others. In a study using a quasi-experimental pre-post design (N = 146), we confirm that the structure of cognitive dispositions of adaptive intervention competence for teaching simulations and mathematical modelling with digital tools can be described empirically by a two-dimensional model. In addition, the effectiveness of the course is examined and confirmed quantitatively. Finally, the results are discussed, especially against the background of the sample and the research design, and conclusions are derived for possibilities of promoting professional adaptive intervention competence in university courses.
For a connected real Lie group G we consider the canonical standard-ordered star product arising from the canonical global symbol calculus based on the half-commutator connection of G. This star product trivially converges on polynomial functions on T\(^*\)G thanks to its homogeneity. We define a nuclear Fréchet algebra of certain analytic functions on T\(^*\)G, for which the standard-ordered star product is shown to be a well-defined continuous multiplication, depending holomorphically on the deformation parameter \(\hbar\). This nuclear Fréchet algebra is realized as the completed (projective) tensor product of a nuclear Fréchet algebra of entire functions on G with an appropriate nuclear Fréchet algebra of functions on \({\mathfrak {g}}^*\). The passage to the Weyl-ordered star product, i.e. the Gutt star product on T\(^*\)G, is shown to preserve this function space, yielding the continuity of the Gutt star product with holomorphic dependence on \(\hbar\).
Let (ϕ\(_t\))\(_{t≥0}\) be a semigroup of holomorphic functions in the unit disk \(\mathbb {D}\) and K a compact subset of \(\mathbb {D}\). We investigate the conditions under which the backward orbit of K under the semigroup exists. Subsequently, the geometric characteristics, as well as, potential theoretic quantities for the backward orbit of K are examined. More specifically, results are obtained concerning the asymptotic behavior of its hyperbolic area and diameter, the harmonic measure and the capacity of the condenser that K forms with the unit disk.
This paper studies differential graded modules and representations up to homotopy of Lie n-algebroids, for general \(n\in {\mathbb {N}}\). The adjoint and coadjoint modules are described, and the corresponding split versions of the adjoint and coadjoint representations up to homotopy are explained. In particular, the case of Lie 2-algebroids is analysed in detail. The compatibility of a Poisson bracket with the homological vector field of a Lie n-algebroid is shown to be equivalent to a morphism from the coadjoint module to the adjoint module, leading to an alternative characterisation of non-degeneracy of higher Poisson structures. Moreover, the Weil algebra of a Lie n-algebroid is computed explicitly in terms of splittings, and representations up to homotopy of Lie n-algebroids are used to encode decomposed VB-Lie n-algebroid structures on double vector bundles.
We extend Bourgain’s bound for the order of growth of the Riemann zeta function on the critical line to Lerch zeta functions. More precisely, we prove L(λ, α, 1/2 + it) ≪ t\(^{13/84+ϵ}\) as t → ∞. For both, the Riemann zeta function as well as for the more general Lerch zeta function, it is conjectured that the right-hand side can be replaced by t\(^ϵ\) (which is the so-called Lindelöf hypothesis). The growth of an analytic function is closely related to the distribution of its zeros.
We give a collection of 16 examples which show that compositions \(g\) \(\circ\) \(f\) of well-behaved functions \(f\) and \(g\) can be badly behaved. Remarkably, in 10 of the 16 examples it suffices to take as outer function \(g\) simply a power-type or characteristic function. Such a collection of examples may serve as a source of exercises for a calculus course.
For a graph \(\Gamma\) , let K be the smallest field containing all eigenvalues of the adjacency matrix of \(\Gamma\) . The algebraic degree \(\deg (\Gamma )\) is the extension degree \([K:\mathbb {Q}]\). In this paper, we completely determine the algebraic degrees of Cayley graphs over abelian groups and dihedral groups.
Mathematical concepts are regularly used in media reports concerning the Covid-19 pandemic. These include growth models, which attempt to explain or predict the effectiveness of interventions and developments, as well as the reproductive factor. Our contribution has the aim of showing that basic mental models about exponential growth are important for understanding media reports of Covid-19. Furthermore, we highlight how the coronavirus pandemic can be used as a context in mathematics classrooms to help students understand that they can and should question media reports on their own, using their mathematical knowledge. Therefore, we first present the role of mathematical modelling in achieving these goals in general. The same relevance applies to the necessary basic mental models of exponential growth. Following this description, based on three topics, namely, investigating the type of growth, questioning given course models, and determining exponential factors at different times, we show how the presented theoretical aspects manifest themselves in teaching examples when students are given the task of reflecting critically on existing media reports. Finally, the value of the three topics regarding the intended goals is discussed and conclusions concerning the possibilities and limits of their use in schools are drawn.
We generalize a theorem by Titchmarsh about the mean value of Hardy’s \(Z\)-function at the Gram points to the Hecke \(L\)-functions, which in turn implies the weak Gram law for them. Instead of proceeding analogously to Titchmarsh with an approximate functional equation we employ a different method using contour integration.
The concept of derivative is characterised with reference to four basic mental models. These are described as theoretical constructs based on theoretical considerations. The four basic mental models—local rate of change, tangent slope, local linearity and amplification factor—are not only quantified empirically but are also validated. To this end, a test instrument for measuring students’ characteristics of basic mental models is presented and analysed regarding quality criteria.
Mathematics students (n = 266) were tested with this instrument. The test results show that the four basic mental models of the derivative can be reconstructed among the students with different characteristics. The tangent slope has the highest agreement values across all tasks. The agreement on explanations based on the basic mental model of rate of change is not as strongly established among students as one would expect due to framework settings in the school system by means of curricula and educational standards. The basic mental model of local linearity plays a rather subordinate role. The amplification factor achieves the lowest agreement values. In addition, cluster analysis was conducted to identify different subgroups of the student population. Moreover, the test results can be attributed to characteristics of the task types as well as to the students’ previous experiences from mathematics classes by means of qualitative interpretation. These and other results of students’ basic mental models of the derivative are presented and discussed in detail.
Composite optimization problems, where the sum of a smooth and a merely lower semicontinuous function has to be minimized, are often tackled numerically by means of proximal gradient methods as soon as the lower semicontinuous part of the objective function is of simple enough structure. The available convergence theory associated with these methods (mostly) requires the derivative of the smooth part of the objective function to be (globally) Lipschitz continuous, and this might be a restrictive assumption in some practically relevant scenarios. In this paper, we readdress this classical topic and provide convergence results for the classical (monotone) proximal gradient method and one of its nonmonotone extensions which are applicable in the absence of (strong) Lipschitz assumptions. This is possible since, for the price of forgoing convergence rates, we omit the use of descent-type lemmas in our analysis.
Nowadays, science, technology, engineering, and mathematics (STEM) play a critical role in a nation’s global competitiveness and prosperity. Thus, there is a need to educate students in these subjects to meet the current and future demands of personal life and society. While applications, especially in science, engineering, and technology, are directly obvious, mathematics underpins the other STEM disciplines. It is recognized that mathematics is the foundation for all other STEM disciplines; the role of mathematics in classrooms is not clear yet. Therefore, the question arises: What is the current role of mathematics in secondary STEM classrooms? To answer this question, we conducted a systematic literature review based on three publication databases (Web of Science, ERIC, and EBSCO Teacher Referral Center). This literature review paper is intended to contribute to the current state of the role of mathematics in STEM education in secondary classrooms. Through the search, starting with 1910 documents, only 14 eligible documents were found. In these, mathematics is often seen as a minor matter and a means to an end in the eyes of science educators. From this, we conclude that the role of mathematics in the STEM classroom should be further strengthened. Overall, the paper highlights a major research gap, and proposes possible initial solutions to close it.
Ó. Blasco and S. Pott showed that the supremum of operator norms over L\(^{2}\) of all bicommutators (with the same symbol) of one-parameter Haar multipliers dominates the biparameter dyadic product BMO norm of the symbol itself. In the present work we extend this result to the Bloom setting, and to any exponent 1 < p < ∞. The main tool is a new characterization in terms of paraproducts and two-weight John–Nirenberg inequalities for dyadic product BMO in the Bloom setting. We also extend our results to the whole scale of indexed spaces between little bmo and product BMO in the general multiparameter setting, with the appropriate iterated commutator in each case.
Bivariate copula monitoring
(2022)
The assumption of multivariate normality underlying the Hotelling T\(^{2}\) chart is often violated for process data. The multivariate dependency structure can be separated from marginals with the help of copula theory, which permits to model association structures beyond the covariance matrix. Copula‐based estimation and testing routines have reached maturity regarding a variety of practical applications. We have constructed a rich design matrix for the comparison of the Hotelling T\(^{2}\) chart with the copula test by Verdier and the copula test by Vuong, which allows for weighting the observations adaptively. Based on the design matrix, we have conducted a large and computationally intensive simulation study. The results show that the copula test by Verdier performs better than Hotelling T\(^{2}\) in a large variety of out‐of‐control cases, whereas the weighted Vuong scheme often fails to provide an improvement.
In financial mathematics, it is a typical approach to approximate financial markets operating in discrete time by continuous-time models such as the Black Scholes model. Fitting this model gives rise to difficulties due to the discrete nature of market data. We thus model the pricing process of financial derivatives by the Black Scholes equation, where the volatility is a function of a finite number of random variables. This reflects an influence of uncertain factors when determining volatility. The aim is to quantify the effect of this uncertainty when computing the price of derivatives. Our underlying method is the generalized Polynomial Chaos (gPC) method in order to numerically compute the uncertainty of the solution by the stochastic Galerkin approach and a finite difference method. We present an efficient numerical variation of this method, which is based on a machine learning technique, the so-called Bi-Fidelity approach. This is illustrated with numerical examples.
The aim of this work is to provide further insight into the qualitative behavior of mechanical systems that are well described by Lennard-Jones type interactions on an atomistic scale. By means of Gamma-convergence techniques, we study the continuum limit of one-dimensional chains of atoms with finite range interactions of Lennard-Jones type, including the classical Lennard-Jones potentials. So far, explicit formula for the continuum limit were only available for the case of nearest and next-to-nearest neighbour interactions. In this work, we provide an explicit expression for the continuum limit in the case of finite range interactions. The obtained homogenization formula is given by the convexification of a Cauchy-Born energy density. Furthermore, we study rescaled energies in which bulk and surface contributions scale in the same way. The related discrete-to-continuum limit yields a rigorous derivation of a one-dimensional version of Griffith' fracture energy and thus generalizes earlier derivations for nearest and next-to-nearest neighbors to the case of finite range interactions. A crucial ingredient to our proofs is a novel decomposition of the energy that allows for re fined estimates.
In this paper, we prove an asymptotic formula for the sum of the values of the periodic zeta-function at the nontrivial zeros of the Riemann zeta-function (up to some height) which are symmetrical on the real line and the critical line. This is an extension of the previous results due to Garunkštis, Kalpokas, and, more recently, Sowa. Whereas Sowa's approach was assuming the yet unproved Riemann hypothesis, our result holds unconditionally.
We consider the Bathnagar–Gross–Krook (BGK) model, an approximation of the Boltzmann equation, describing the time evolution of a single momoatomic rarefied gas and satisfying the same two main properties (conservation properties and entropy inequality). However, in practical applications, one often has to deal with two additional physical issues. First, a gas often does not consist of only one species, but it consists of a mixture of different species. Second, the particles can store energy not only in translational degrees of freedom but also in internal degrees of freedom such as rotations or vibrations (polyatomic molecules). Therefore, here, we will present recent BGK models for gas mixtures for mono- and polyatomic particles and the existing mathematical theory for these models.