Identification of unknown parameters and prediction with hierarchical matrice...Alexander Litvinenko
We compare four numerical methods for the prediction of missing values in four different datasets.
These methods are 1) the hierarchical maximum likelihood estimation (H-MLE), and three machine learning (ML) methods, which include 2) k-nearest neighbors (kNN), 3) random forest, and 4) Deep Neural Network (DNN).
From the ML methods, the best results (for considered datasets) were obtained by the kNN method with three (or seven) neighbors.
On one dataset, the MLE method showed a smaller error than the kNN method, whereas, on another, the kNN method was better.
The MLE method requires a lot of linear algebra computations and works fine on almost all datasets. Its result can be improved by taking a smaller threshold and more accurate hierarchical matrix arithmetics. To our surprise, the well-known kNN method produces similar results as H-MLE and worked much faster.
CAPE PURE MATHEMATICS UNIT 2 MODULE 1 PRACTICE QUESTIONSCarlon Baird
dy/dx = (x - 3y)/(6x - 4)
The stationary points on the curve C occur when tan(x) = 2.
The equation of the tangent to C at the point where x=0 is y = 2ex.
CAPE PURE MATHEMATICS UNIT 2 MODULE 2 PRACTICE QUESTIONSCarlon Baird
This document contains practice questions on sequences, series, and approximations from a CAPE Pure Mathematics unit. Question 1 covers finding terms of sequences defined recursively and evaluating finite sums. Question 2 involves finding expressions for terms of sequences defined recursively and finding their sums. Later questions cover topics like proving identities using induction, evaluating infinite series, approximating functions using Taylor series, and finding roots of equations numerically. The questions provide worked examples of key concepts in sequences, series, and approximations.
Identification of the Mathematical Models of Complex Relaxation Processes in ...Vladimir Bakhrushin
The approach to solving the problem of complex relaxation spectra is presented.
Presentation for the XI International Conference on Defect interaction and anelastic phenomena in solids. Tula, 2007.
Quantum mechanics 1st edition mc intyre solutions manualSelina333
Quantum Mechanics 1st Edition McIntyre Solutions Manual
Download at: https://goo.gl/SdC7Ef
quantum mechanics david mcintyre solutions pdf
quantum mechanics mcintyre pdf
quantum mechanics a paradigms approach solutions pdf
quantum mechanics mcintyre solutions pdf
quantum mechanics a paradigms approach solution manual
quantum mechanics mcintyre solutions manual pdf
hidden life of prayer
Trial pahang 2014 spm add math k2 dan skema [scan]Cikgu Pejal
This document contains formulae and tables that may be helpful for answering questions in the Additional Mathematics paper. The first section lists common algebraic, calculus, statistics, geometry, trigonometry and normal distribution formulae. The second section contains a table with values of the standard normal distribution function Φ(z) for both positive and negative z-values up to z=3. The document then lists the questions in Section A of the paper. Section A contains 3 questions, the first two involve solving simultaneous equations and finding terms in a geometric progression, and the third finds the smallest value of n such that the nth term is less than 0.01.
Bellman-Ford-Moore Algorithm and Dijkstra’s AlgorithmFulvio Corno
Bellman-Ford-Moore Algorithm and Dijkstra’s Algorithm
Teaching material for the course of "Tecniche di Programmazione" at Politecnico di Torino in year 2012/2013. More information: http://bit.ly/tecn-progr
1. The graph of the equation y = x^2 - 1 can be viewed as a torus when considered with points at infinity and over the complex numbers.
2. Intersections of the graph with planes where x or y is set to a constant reveal curves hinting at the toric structure, such as ellipses or lemniscates.
3. When extended to projective space over the complex numbers, the graph can be seen topologically as a two-dimensional surface in four-dimensional space, matching the geometry of a torus.
"SSumM: Sparse Summarization of Massive Graphs", KDD 2020KyuhanLee4
A presentation slides of Kyuhan Lee, Hyeonsoo Jo, Jihoon Ko, Sungsu Lim, Kijung Shin, "SSumM: Sparse Summarization of Massive Graphs", KDD 2020.
Given a graph G and the desired size k in bits, how can we summarize G within k bits, while minimizing the information loss?
Large-scale graphs have become omnipresent, posing considerable computational challenges. Analyzing such large graphs can be fast and easy if they are compressed sufficiently to fit in main memory or even cache. Graph summarization, which yields a coarse-grained summary graph with merged nodes, stands out with several advantages among graph compression techniques. Thus, a number of algorithms have been developed for obtaining a concise summary graph with little information loss or equivalently small reconstruction error. However, the existing methods focus solely on reducing the number of nodes, and they often yield dense summary graphs, failing to achieve better compression rates. Moreover, due to their limited scalability, they can be applied only to moderate-size graphs.
In this work, we propose SSumM, a scalable and effective graph-summarization algorithm that yields a sparse summary graph. SSumM not only merges nodes together but also sparsifies the summary graph, and the two strategies are carefully balanced based on the minimum description length principle. Compared with state-of-the-art competitors, SSumM is (a) Concise: yields up to 11.2X smaller summary graphs with similar reconstruction error, (b) Accurate: achieves up to 4.2X smaller reconstruction error with similarly concise outputs, and (c) Scalable: summarizes 26X larger graphs while exhibiting linear scalability. We validate these advantages through extensive experiments on 10 real-world graphs.
Second-order Cosmological Perturbations Engendered by Point-like MassesMaxim Eingorn
R. Brilenkov and M. Eingorn, Second-order cosmological perturbations engendered by point-like masses, ApJ 845 (2017) 153: http://iopscience.iop.org/article/10.3847/1538-4357/aa81cd
In the ΛCDM framework, presenting nonrelativistic matter inhomogeneities as discrete massive particles, we develop the second‐order cosmological perturbation theory. Our approach relies on the weak gravitational field limit. The derived equations for the second‐order scalar, vector, and tensor metric corrections are suitable at arbitrary distances, including regions with nonlinear contrasts of the matter density. We thoroughly verify fulfillment of all Einstein equations, as well as self‐consistency of order assignments. In addition, we achieve logical positive results in the Minkowski background limit. Feasible investigations of the cosmological backreaction manifestations by means of relativistic simulations are also outlined.
This document discusses numerical methods for solving ordinary differential equations (ODEs), including Euler methods, Runge-Kutta methods, and their application to problems in math and physics. It provides examples of using Euler's method and Runge-Kutta methods to solve ODEs describing a block-spring system. Code implementations of the Euler and Runge-Kutta methods are presented along with comparisons of numerical solutions to exact solutions.
"Incremental Lossless Graph Summarization", KDD 2020지훈 고
A presentation slides of Jihoon Ko*, Yunbum Kook* and Kijung Shin, "Incremental Lossless Graph Summarization", KDD 2020.
Given a fully dynamic graph, represented as a stream of edge insertions and deletions, how can we obtain and incrementally update a lossless summary of its current snapshot?
As large-scale graphs are prevalent, concisely representing them is inevitable for efficient storage and analysis. Lossless graph summarization is an effective graph-compression technique with many desirable properties. It aims to compactly represent the input graph as (a) a summary graph consisting of supernodes (i.e., sets of nodes) and superedges (i.e., edges between supernodes), which provide a rough description, and (b) edge corrections which fix errors induced by the rough description. While a number of batch algorithms, suited for static graphs, have been developed for rapid and compact graph summarization, they are highly inefficient in terms of time and space for dynamic graphs, which are common in practice.
In this work, we propose MoSSo, the first incremental algorithm for lossless summarization of fully dynamic graphs. In response to each change in the input graph, MoSSo updates the output representation by repeatedly moving nodes among supernodes. MoSSo decides nodes to be moved and their destinations carefully but rapidly based on several novel ideas. Through extensive experiments on 10 real graphs, we show MoSSo is (a) Fast and 'any time': processing each change in near-constant time (less than 0.1 millisecond), up to 7 orders of magnitude faster than running state-of-the-art batch methods, (b) Scalable: summarizing graphs with hundreds of millions of edges, requiring sub-linear memory during the process, and (c) Effective: achieving comparable compression ratios even to state-of-the-art batch methods.
The document discusses hypergraph motifs, which describe connectivity patterns between three connected hyperedges in a hypergraph. It proposes MoCHy, a family of parallel algorithms for counting instances of hypergraph motifs in large hypergraphs. Experimental results on real-world hypergraphs from different domains show that their motif distributions differ significantly from randomized hypergraphs, and MoCHy can efficiently count motifs in large hypergraphs.
E. Canay and M. Eingorn
Physics of the Dark Universe 29 (2020) 100565
DOI: 10.1016/j.dark.2020.100565
https://authors.elsevier.com/a/1aydL7t6qq5DB0
https://arxiv.org/abs/2002.00437
Two distinct perturbative approaches have been recently formulated within General Relativity, arguing for the screening of gravity in the ΛCDM Universe. We compare them and show that the offered screening concepts, each characterized by its own interaction range, can peacefully coexist. Accordingly, we advance a united scheme, determining the gravitational potential at all scales, including regions of nonlinear density contrasts, by means of a simple Helmholtz equation with the effective cosmological screening length. In addition, we claim that cosmic structures may not grow at distances above this Yukawa range and confront its current value with dimensions of the largest known objects in the Universe.
This document provides a review of exercises for a Math 112 final exam. It contains 31 multi-part exercises covering topics like graphing, logarithms, trigonometry, and word problems. The review is intended to help students practice problems similar to what may appear on the exam. The exam will have two parts, one allowing a calculator and one not.
This document provides model solutions to questions from the JEE Advanced 2013 exam. It includes solutions to multiple choice and numerical questions across physics, chemistry and mathematics. Some key details include:
- Solutions to 20 multiple choice questions from Paper I Section I and II of the exam.
- Solutions involving concepts like circular motion, momentum, electrostatics, thermodynamics and more.
- Solutions to 4 numerical questions involving calculations related to decay of particles, fundamental frequency of a vibrating string, and more.
- Solutions to 20 multiple choice questions from Paper II Section I, II and III covering topics in chemistry, organic chemistry, coordination chemistry and biochemistry.
- Solutions to 4 numerical questions related
This document describes a code structure for calculating and visualizing electric potential and field from point charges. It discusses:
1) Calculating the potential and electric field at grid points due to multiple point charges using superposition principles.
2) Interpolating sparse potential data to generate smooth 2D potential maps.
3) Representing the electric field as vectors showing position, magnitude, and direction originating from point charges.
The code reads charge and position inputs, calculates potentials and fields on a grid, interpolates the potential data, and outputs files to generate vector maps visualizing the electric potential and field.
1. The document discusses numerical integration techniques for approximating definite integrals that cannot be solved analytically. It covers basic techniques like the rectangle, midpoint, and trapezoid rules as well as more accurate techniques like Simpson's rule.
2. Examples are provided to demonstrate calculating definite integrals numerically to approximate values like the natural logarithm of numbers. The document also introduces Monte Carlo integration techniques using random sampling.
3. As an example problem, the document calculates the final speed of a box moving under a time-varying force using numerical integration over the integral expression for work. The Simpson's rule is identified as an approach to implement in a programming code to solve this example.
This document discusses different types of kinematic synthesis including function generation, path generation, and motion generation. It then focuses on precision point synthesis for two-position and three-position synthesis. For two-position synthesis, the document outlines solving the vector loop equation to find the unknown link lengths and angles given some known precision points, coupler angle, and link lengths. For three-position synthesis, two vector loop equations are written and solved similarly. The document compares the graphical and analytical methods and discusses an example of three-position synthesis with specified fixed pivots.
This document appears to be the table of contents and problems from Chapter 0 of a mathematics textbook. The table of contents lists 17 chapters and their corresponding page numbers. The problems cover a range of algebra topics including integers, rational numbers, properties of operations, solving equations, and rational expressions. There are over 70 problems presented without solutions for students to work through.
This document discusses working with rational exponents. It introduces the rational exponent theorem, which states that for any nonnegative real number x and positive integers m and n, x^(m/n) = (x^(1/n))^m, the mth power of the nth root of x. Several examples are worked through to demonstrate simplifying expressions with rational exponents using this theorem. Special cases for negative rational exponents are also discussed. Homework problems are assigned from the textbook.
The document provides solutions to questions from an IIT-JEE mathematics exam. It includes 8 questions worth 2 marks each, 8 questions worth 4 marks each, and 2 questions worth 6 marks each. The solutions solve problems related to probability, trigonometry, geometry, calculus, and loci. The summary focuses on the high-level structure and content of the document.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
Communication systems solution manual 5th editionTayeen Ahmed
This document contains solutions to problems from Communications Systems, 5th edition. It includes solutions for time domain representations of pulses, their frequency domain counterparts obtained using Fourier transforms, properties of even and odd functions, and bounds on the bandwidth of time-limited signals. The solutions demonstrate properties like linearity, time-shifting, differentiation in the time and frequency domains, and the relationships between a function and its Fourier transform.
This document is a solutions manual for a textbook on communication systems. It provides step-by-step solutions to problems from each chapter of the textbook. The problems cover topics such as signal representations using Fourier series and integrals, power calculations for periodic signals, and bandpass signal representations. The solutions demonstrate techniques for analyzing and working with signals commonly encountered in electrical communication systems.
Identification of unknown parameters and prediction of missing values. Compar...Alexander Litvinenko
H-matrix approximation of large Mat\'{e}rn covariance matrices, Gaussian log-likelihoods.
Identifying unknown parameters and making predictions
Comparison with machine learning methods.
kNN is easy to implement and shows promising results.
Application of Parallel Hierarchical Matrices in Spatial Statistics and Param...Alexander Litvinenko
Part 1: Parallel H-matrices in spatial statistics
1. Motivation: improve statistical model
2. Tools: Hierarchical matrices [Hackbusch 1999]
3. Matern covariance function and joint Gaussian likelihood
4. Identification of unknown parameters via maximizing Gaussian
log-likelihood
5. Implementation with HLIBPro.
Bellman-Ford-Moore Algorithm and Dijkstra’s AlgorithmFulvio Corno
Bellman-Ford-Moore Algorithm and Dijkstra’s Algorithm
Teaching material for the course of "Tecniche di Programmazione" at Politecnico di Torino in year 2012/2013. More information: http://bit.ly/tecn-progr
1. The graph of the equation y = x^2 - 1 can be viewed as a torus when considered with points at infinity and over the complex numbers.
2. Intersections of the graph with planes where x or y is set to a constant reveal curves hinting at the toric structure, such as ellipses or lemniscates.
3. When extended to projective space over the complex numbers, the graph can be seen topologically as a two-dimensional surface in four-dimensional space, matching the geometry of a torus.
"SSumM: Sparse Summarization of Massive Graphs", KDD 2020KyuhanLee4
A presentation slides of Kyuhan Lee, Hyeonsoo Jo, Jihoon Ko, Sungsu Lim, Kijung Shin, "SSumM: Sparse Summarization of Massive Graphs", KDD 2020.
Given a graph G and the desired size k in bits, how can we summarize G within k bits, while minimizing the information loss?
Large-scale graphs have become omnipresent, posing considerable computational challenges. Analyzing such large graphs can be fast and easy if they are compressed sufficiently to fit in main memory or even cache. Graph summarization, which yields a coarse-grained summary graph with merged nodes, stands out with several advantages among graph compression techniques. Thus, a number of algorithms have been developed for obtaining a concise summary graph with little information loss or equivalently small reconstruction error. However, the existing methods focus solely on reducing the number of nodes, and they often yield dense summary graphs, failing to achieve better compression rates. Moreover, due to their limited scalability, they can be applied only to moderate-size graphs.
In this work, we propose SSumM, a scalable and effective graph-summarization algorithm that yields a sparse summary graph. SSumM not only merges nodes together but also sparsifies the summary graph, and the two strategies are carefully balanced based on the minimum description length principle. Compared with state-of-the-art competitors, SSumM is (a) Concise: yields up to 11.2X smaller summary graphs with similar reconstruction error, (b) Accurate: achieves up to 4.2X smaller reconstruction error with similarly concise outputs, and (c) Scalable: summarizes 26X larger graphs while exhibiting linear scalability. We validate these advantages through extensive experiments on 10 real-world graphs.
Second-order Cosmological Perturbations Engendered by Point-like MassesMaxim Eingorn
R. Brilenkov and M. Eingorn, Second-order cosmological perturbations engendered by point-like masses, ApJ 845 (2017) 153: http://iopscience.iop.org/article/10.3847/1538-4357/aa81cd
In the ΛCDM framework, presenting nonrelativistic matter inhomogeneities as discrete massive particles, we develop the second‐order cosmological perturbation theory. Our approach relies on the weak gravitational field limit. The derived equations for the second‐order scalar, vector, and tensor metric corrections are suitable at arbitrary distances, including regions with nonlinear contrasts of the matter density. We thoroughly verify fulfillment of all Einstein equations, as well as self‐consistency of order assignments. In addition, we achieve logical positive results in the Minkowski background limit. Feasible investigations of the cosmological backreaction manifestations by means of relativistic simulations are also outlined.
This document discusses numerical methods for solving ordinary differential equations (ODEs), including Euler methods, Runge-Kutta methods, and their application to problems in math and physics. It provides examples of using Euler's method and Runge-Kutta methods to solve ODEs describing a block-spring system. Code implementations of the Euler and Runge-Kutta methods are presented along with comparisons of numerical solutions to exact solutions.
"Incremental Lossless Graph Summarization", KDD 2020지훈 고
A presentation slides of Jihoon Ko*, Yunbum Kook* and Kijung Shin, "Incremental Lossless Graph Summarization", KDD 2020.
Given a fully dynamic graph, represented as a stream of edge insertions and deletions, how can we obtain and incrementally update a lossless summary of its current snapshot?
As large-scale graphs are prevalent, concisely representing them is inevitable for efficient storage and analysis. Lossless graph summarization is an effective graph-compression technique with many desirable properties. It aims to compactly represent the input graph as (a) a summary graph consisting of supernodes (i.e., sets of nodes) and superedges (i.e., edges between supernodes), which provide a rough description, and (b) edge corrections which fix errors induced by the rough description. While a number of batch algorithms, suited for static graphs, have been developed for rapid and compact graph summarization, they are highly inefficient in terms of time and space for dynamic graphs, which are common in practice.
In this work, we propose MoSSo, the first incremental algorithm for lossless summarization of fully dynamic graphs. In response to each change in the input graph, MoSSo updates the output representation by repeatedly moving nodes among supernodes. MoSSo decides nodes to be moved and their destinations carefully but rapidly based on several novel ideas. Through extensive experiments on 10 real graphs, we show MoSSo is (a) Fast and 'any time': processing each change in near-constant time (less than 0.1 millisecond), up to 7 orders of magnitude faster than running state-of-the-art batch methods, (b) Scalable: summarizing graphs with hundreds of millions of edges, requiring sub-linear memory during the process, and (c) Effective: achieving comparable compression ratios even to state-of-the-art batch methods.
The document discusses hypergraph motifs, which describe connectivity patterns between three connected hyperedges in a hypergraph. It proposes MoCHy, a family of parallel algorithms for counting instances of hypergraph motifs in large hypergraphs. Experimental results on real-world hypergraphs from different domains show that their motif distributions differ significantly from randomized hypergraphs, and MoCHy can efficiently count motifs in large hypergraphs.
E. Canay and M. Eingorn
Physics of the Dark Universe 29 (2020) 100565
DOI: 10.1016/j.dark.2020.100565
https://authors.elsevier.com/a/1aydL7t6qq5DB0
https://arxiv.org/abs/2002.00437
Two distinct perturbative approaches have been recently formulated within General Relativity, arguing for the screening of gravity in the ΛCDM Universe. We compare them and show that the offered screening concepts, each characterized by its own interaction range, can peacefully coexist. Accordingly, we advance a united scheme, determining the gravitational potential at all scales, including regions of nonlinear density contrasts, by means of a simple Helmholtz equation with the effective cosmological screening length. In addition, we claim that cosmic structures may not grow at distances above this Yukawa range and confront its current value with dimensions of the largest known objects in the Universe.
This document provides a review of exercises for a Math 112 final exam. It contains 31 multi-part exercises covering topics like graphing, logarithms, trigonometry, and word problems. The review is intended to help students practice problems similar to what may appear on the exam. The exam will have two parts, one allowing a calculator and one not.
This document provides model solutions to questions from the JEE Advanced 2013 exam. It includes solutions to multiple choice and numerical questions across physics, chemistry and mathematics. Some key details include:
- Solutions to 20 multiple choice questions from Paper I Section I and II of the exam.
- Solutions involving concepts like circular motion, momentum, electrostatics, thermodynamics and more.
- Solutions to 4 numerical questions involving calculations related to decay of particles, fundamental frequency of a vibrating string, and more.
- Solutions to 20 multiple choice questions from Paper II Section I, II and III covering topics in chemistry, organic chemistry, coordination chemistry and biochemistry.
- Solutions to 4 numerical questions related
This document describes a code structure for calculating and visualizing electric potential and field from point charges. It discusses:
1) Calculating the potential and electric field at grid points due to multiple point charges using superposition principles.
2) Interpolating sparse potential data to generate smooth 2D potential maps.
3) Representing the electric field as vectors showing position, magnitude, and direction originating from point charges.
The code reads charge and position inputs, calculates potentials and fields on a grid, interpolates the potential data, and outputs files to generate vector maps visualizing the electric potential and field.
1. The document discusses numerical integration techniques for approximating definite integrals that cannot be solved analytically. It covers basic techniques like the rectangle, midpoint, and trapezoid rules as well as more accurate techniques like Simpson's rule.
2. Examples are provided to demonstrate calculating definite integrals numerically to approximate values like the natural logarithm of numbers. The document also introduces Monte Carlo integration techniques using random sampling.
3. As an example problem, the document calculates the final speed of a box moving under a time-varying force using numerical integration over the integral expression for work. The Simpson's rule is identified as an approach to implement in a programming code to solve this example.
This document discusses different types of kinematic synthesis including function generation, path generation, and motion generation. It then focuses on precision point synthesis for two-position and three-position synthesis. For two-position synthesis, the document outlines solving the vector loop equation to find the unknown link lengths and angles given some known precision points, coupler angle, and link lengths. For three-position synthesis, two vector loop equations are written and solved similarly. The document compares the graphical and analytical methods and discusses an example of three-position synthesis with specified fixed pivots.
This document appears to be the table of contents and problems from Chapter 0 of a mathematics textbook. The table of contents lists 17 chapters and their corresponding page numbers. The problems cover a range of algebra topics including integers, rational numbers, properties of operations, solving equations, and rational expressions. There are over 70 problems presented without solutions for students to work through.
This document discusses working with rational exponents. It introduces the rational exponent theorem, which states that for any nonnegative real number x and positive integers m and n, x^(m/n) = (x^(1/n))^m, the mth power of the nth root of x. Several examples are worked through to demonstrate simplifying expressions with rational exponents using this theorem. Special cases for negative rational exponents are also discussed. Homework problems are assigned from the textbook.
The document provides solutions to questions from an IIT-JEE mathematics exam. It includes 8 questions worth 2 marks each, 8 questions worth 4 marks each, and 2 questions worth 6 marks each. The solutions solve problems related to probability, trigonometry, geometry, calculus, and loci. The summary focuses on the high-level structure and content of the document.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
Communication systems solution manual 5th editionTayeen Ahmed
This document contains solutions to problems from Communications Systems, 5th edition. It includes solutions for time domain representations of pulses, their frequency domain counterparts obtained using Fourier transforms, properties of even and odd functions, and bounds on the bandwidth of time-limited signals. The solutions demonstrate properties like linearity, time-shifting, differentiation in the time and frequency domains, and the relationships between a function and its Fourier transform.
This document is a solutions manual for a textbook on communication systems. It provides step-by-step solutions to problems from each chapter of the textbook. The problems cover topics such as signal representations using Fourier series and integrals, power calculations for periodic signals, and bandpass signal representations. The solutions demonstrate techniques for analyzing and working with signals commonly encountered in electrical communication systems.
Identification of unknown parameters and prediction of missing values. Compar...Alexander Litvinenko
H-matrix approximation of large Mat\'{e}rn covariance matrices, Gaussian log-likelihoods.
Identifying unknown parameters and making predictions
Comparison with machine learning methods.
kNN is easy to implement and shows promising results.
Application of Parallel Hierarchical Matrices in Spatial Statistics and Param...Alexander Litvinenko
Part 1: Parallel H-matrices in spatial statistics
1. Motivation: improve statistical model
2. Tools: Hierarchical matrices [Hackbusch 1999]
3. Matern covariance function and joint Gaussian likelihood
4. Identification of unknown parameters via maximizing Gaussian
log-likelihood
5. Implementation with HLIBPro.
Hierarchical matrix approximation of large covariance matricesAlexander Litvinenko
We research class of Matern covariance matrices and their approximability in the H-matrix format. Further tasks are compute H-Cholesky factorization, trace, determinant, quadratic form, loglikelihood. Later H-matrices can be applied in kriging.
Approximation of large Matern covariance functions in the H-matrix format. We computed relative errors in spectral, Frobenius norms as well as the Kullback-Leibler divergence. Storage and computational costs are drastically reduced.
Hierarchical matrix techniques for maximum likelihood covariance estimationAlexander Litvinenko
1. We apply hierarchical matrix techniques (HLIB, hlibpro) to approximate huge covariance matrices. We are able to work with 250K-350K non-regular grid nodes.
2. We maximize a non-linear, non-convex Gaussian log-likelihood function to identify hyper-parameters of covariance.
This document contains 4 math exercises involving matrix operations and linear algebra calculations. Exercise 1 involves multiplying and adding matrices. Exercise 2 calculates determinants and inverses of matrices. Exercise 3 solves systems of linear equations. Exercise 4 shows that the determinant of AB is equal to the determinant of A times the determinant of B.
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practicesPerform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform good laboratory practices
Perform g
My presentation at University of Nottingham "Fast low-rank methods for solvin...Alexander Litvinenko
Overview of my (with co-authors) low-rank tensor methods for solving PDEs with uncertain coefficients. Connection with Bayesian Update. Solving a coupled system: stochastic forward and stochastic inverse.
Numerical Methods: curve fitting and interpolationNikolai Priezjev
This document discusses curve fitting and interpolation techniques. It covers linear regression using the least squares method to fit data to a straight line model. It also discusses fitting data to other functions like exponential, logarithmic and polynomial models. For polynomial regression, a second order polynomial is presented which requires solving a system of equations to determine the coefficients that minimize the residual errors between measured and modeled data. An example demonstrates applying these methods to find the coefficients for a straight line and second order polynomial fit to sample data sets.
This document discusses statistical process control and control charts. It defines the goals of control charts as collecting and visually presenting data to see when trends or out-of-control points occur. Process control charts graph sample data over time and show the process average and upper and lower control limits. Attribute control charts indicate whether points are in or out of tolerance, while variables charts measure attributes like length, weight or temperature over time. Examples are provided to illustrate p-charts, R-charts and X-bar charts using hotel luggage delivery time data.
Why mathematics is easy to understand, easy to do, and easy to prove in base 60. The number theory behind the creation story and the beginning of time.
Likelihood approximation with parallel hierarchical matrices for large spatia...Alexander Litvinenko
First, we use hierarchical matrices to approximate large Matern covariance matrices and the loglikelihood. Second, we find a maximum of loglikelihood and estimate 3 unknown parameters (covariance length, smoothness and variance).
This document provides rules and examples for factoring polynomials. It discusses factoring using the greatest common factor (GCF), grouping like terms, and other factoring methods for trinomials, perfect square trinomials, differences of two squares, and sums and differences of two cubes. Specific examples are provided to demonstrate each factoring technique.
This document provides rules and examples for factoring polynomials. It discusses factoring using the greatest common factor (GCF), grouping like terms, and other factoring methods for trinomials, perfect square trinomials, differences of two squares, and sums and differences of two cubes. Specific examples are provided to demonstrate each factoring technique.
The treatment of large structural systems may be simplified by dividing the system into
smaller systems called components. The components are related through the
displacement, and force conditions at their junction points. Each component is represented
by mode shapes (or functions).
The goal is to improve parameter estimation for daily moisture field in Mississippi basin. We use maximum likelihood method to estimate unknown parameters of Matern covariance. To speed up computations we apply parallel hierarchical matrices.
Large data sets result large dense matrices, say with 2.000.000 rows and columns. How to work with such large matrices? How to approximate them? How to compute log-likelihood? determination? inverse? All answers are in this work.
Poster to be presented at Stochastic Numerics and Statistical Learning: Theory and Applications Workshop 2024, Kaust, Saudi Arabia, https://cemse.kaust.edu.sa/stochnum/events/event/snsl-workshop-2024.
In this work we have considered a setting that mimics the Henry problem \cite{Simpson2003,Simpson04_Henry}, modeling seawater intrusion into a 2D coastal aquifer. The pure water recharge from the ``land side'' resists the salinisation of the aquifer due to the influx of saline water through the ``sea side'', thereby achieving some equilibrium in the salt concentration. In our setting, following \cite{GRILLO2010}, we consider a fracture on the sea side that significantly increases the permeability of the porous medium.
The flow and transport essentially depend on the geological parameters of the porous medium, including the fracture. We investigated the effects of various uncertainties on saltwater intrusion. We assumed uncertainties in the fracture width, the porosity of the bulk medium, its permeability and the pure water recharge from the land side. The porosity and permeability were modeled by random fields, the recharge by a random but periodic intensity and the thickness by a random variable. We calculated the mean and variance of the salt mass fraction, which is also uncertain.
The main question we investigated in this work was how well the MLMC method can be used to compute statistics of different QoIs. We found that the answer depends on the choice of the QoI. First, not every QoI requires a hierarchy of meshes and MLMC. Second, MLMC requires stable convergence rates for $\EXP{g_{\ell} - g_{\ell-1}}$ and $\Var{g_{\ell} - g_{\ell-1}}$. These rates should be independent of $\ell$. If these convergence rates vary for different $\ell$, then it will be hard to estimate $L$ and $m_{\ell}$, and MLMC will either not work or be suboptimal. We were not able to get stable convergence rates for all levels $\ell=1,\ldots,5$ when the QoI was an integral as in \eqref{eq:integral_box}. We found that for $\ell=1,\ldots 4$ and $\ell=5$ the rate $\alpha$ was different. Further investigation is needed to find the reason for this. Another difficulty is the dependence on time, i.e. the number of levels $L$ and the number of sums $m_{\ell}$ depend on $t$. At the beginning the variability is small, then it increases, and after the process of mixing salt and fresh water has stopped, the variance decreases again.
The number of random samples required at each level was estimated by calculating the decay of the variances and the computational cost for each level. These estimates depend on the minimisation function in the MLMC algorithm.
To achieve the efficiency of the MLMC approach presented in this work, it is essential that the complexity of the numerical solution of each random realisation is proportional to the number of grid vertices on the grid levels.
We investigated the applicability and efficiency of the MLMC approach to the Henry-like problem with uncertain porosity, permeability and recharge. These uncertain parameters were modelled by random fields with three independent random variables. Permeability is a function of porosity. Both functions are time-dependent, have multi-scale behaviour and are defined for two layers. The numerical solution for each random realisation was obtained using the well-known ug4 parallel multigrid solver. The number of random samples required at each level was estimated by calculating the decay of the variances and the computational cost for each level.
The MLMC method was used to compute the expected value and variance of several QoIs, such as the solution at a few preselected points $(t,\bx)$, the solution integrated over a small subdomain, and the time evolution of the freshwater integral. We have found that some QoIs require only 2-3 mesh levels and samples from finer meshes would not significantly improve the result. Other QoIs require more grid levels.
1. Investigated efficiency of MLMC for Henry problem with
uncertain porosity, permeability, and recharge.
2. Uncertainties are modeled by random fields.
3. MLMC could be much faster than MC, 3200 times faster !
4. The time dependence is challenging.
Remarks:
1. Check if MLMC is needed.
2. The optimal number of samples depends on the point (t;x)
3. An advanced MLMC may give better estimates of L and m`.
Density Driven Groundwater Flow with Uncertain Porosity and PermeabilityAlexander Litvinenko
In this work, we solved the density driven groundwater flow problem with uncertain porosity and permeability. An accurate solution of this time-dependent and non-linear problem is impossible because of the presence of natural uncertainties in the reservoir such as porosity and permeability.
Therefore, we estimated the mean value and the variance of the solution, as well as the propagation of uncertainties from the random input parameters to the solution.
We started by defining the Elder-like problem. Then we described the multi-variate polynomial approximation (\gPC) approach and used it to estimate the required statistics of the mass fraction.
Utilizing the \gPC method allowed us
to reduce the computational cost compared to the classical quasi Monte Carlo method.
\gPC assumes that the output function $\sol(t,\bx,\thetab)$ is square-integrable and smooth w.r.t uncertain input variables $\btheta$.
Many factors, such as non-linearity, multiple solutions, multiple stationary states, time dependence and complicated solvers, make the investigation of the convergence of the \gPC method a non-trivial task.
We used an easy-to-implement, but only sub-optimal \gPC technique to quantify the uncertainty. For example, it is known that by increasing the degree of global polynomials (Hermite, Langange and similar), Runge's phenomenon appears. Here, probably local polynomials, splines or their mixtures would be better. Additionally, we used an easy-to-parallelise quadrature rule, which was also only suboptimal. For instance, adaptive choice of sparse grid (or collocation) points \cite{ConradMarzouk13,nobile-sg-mc-2015,Sudret_sparsePCE,CONSTANTINE12,crestaux2009polynomial} would be better, but we were limited by the usage of parallel methods. Adaptive quadrature rules are not (so well) parallelisable. In conclusion, we can report that: a) we developed a highly parallel method to quantify uncertainty in the Elder-like problem; b) with the \gPC of degree 4 we can achieve similar results as with the \QMC method.
In the numerical section we considered two different aquifers - a solid parallelepiped and a solid elliptic cylinder. One of our goals was to see how the domain geometry influences the formation, the number and the shape of fingers.
Since the considered problem is nonlinear,
a high variance in the porosity may result in totally different solutions; for instance, the number of fingers, their intensity and shape, the propagation time, and the velocity may vary considerably.
The number of cells in the presented experiments varied from $241{,}152$ to $15{,}433{,}728$ for the cylindrical domain and from $524{,}288$ to $4{,}194{,}304$ for the parallelepiped. The maximal number of parallel processing units was $600\times 32$, where $600$ is the number of parallel nodes and $32$ is the number of computing cores on each node. The total computing time varied from 2 hours for the coarse mesh to 24 hours for the finest mesh.
Saltwater intrusion occurs when sea levels rise and saltwater moves onto the land. Usually, this occurs during storms, high tides, droughts, or when saltwater penetrates freshwater aquifers and raises the groundwater table. Since groundwater is an essential nutrition and irrigation resource, its salinization may lead to catastrophic consequences. Many acres of farmland may be lost because they can become too wet or salty to grow crops. Therefore, accurate modeling of different scenarios of saline flow is essential to help farmers and researchers develop strategies to improve the soil quality and decrease saltwater intrusion effects.
Saline flow is density-driven and described by a system of time-dependent nonlinear partial differential equations (PDEs). It features convection dominance and can demonstrate very complicated behavior.
As a specific model, we consider a Henry-like problem with uncertain permeability and porosity.
These parameters may strongly affect the flow and transport of salt.
We consider a class of density-driven flow problems. We are particularly interested in the problem of the salinization of coastal aquifers. We consider the Henry saltwater intrusion problem with uncertain porosity, permeability, and recharge parameters as a test case.
The reason for the presence of uncertainties is the lack of knowledge, inaccurate measurements,
and inability to measure parameters at each spatial or time location. This problem is nonlinear and time-dependent. The solution is the salt mass fraction, which is uncertain and changes in time. Uncertainties in porosity, permeability, recharge, and mass fraction are modeled using random fields. This work investigates the applicability of the well-known multilevel Monte Carlo (MLMC) method for such problems. The MLMC method can reduce the total computational and storage costs. Moreover, the MLMC method runs multiple scenarios on different spatial and time meshes and then estimates the mean value of the mass fraction.
The parallelization is performed in both the physical space and stochastic space. To solve every deterministic scenario, we run the parallel multigrid solver ug4 in a black-box fashion.
We use the solution obtained from the quasi-Monte Carlo method as a reference solution.
We investigated the applicability and efficiency of the MLMC approach for the Henry-like problem with uncertain porosity, permeability, and recharge. These uncertain parameters were modeled by random fields with three independent random variables. The numerical solution for each random realization was obtained using the well-known ug4 parallel multigrid solver. The number of required random samples on each level was estimated by computing the decay of the variances and computational costs for each level. We also computed the expected value and variance of the mass fraction in the whole domain, the evolution of the pdfs, the solutions at a few preselected points $(t,\bx)$, and the time evolution of the freshwater integral value. We have found that some QoIs require only 2-3 of the coarsest mesh levels, and samples from finer meshes would not significantly improve the result. Note that a different type of porosity may lead to a different conclusion.
The results show that the MLMC method is faster than the QMC method at the finest mesh. Thus, sampling at different mesh levels makes sense and helps to reduce the overall computational cost.
Here the interest is mainly to compute characterisations like the entropy,
the Kullback-Leibler divergence, more general $f$-divergences, or other such characteristics based on
the probability density. The density is often not available directly,
and it is a computational challenge to just represent it in a numerically
feasible fashion in case the dimension is even moderately large. It
is an even stronger numerical challenge to then actually compute said characteristics
in the high-dimensional case.
The task considered here was the numerical computation of characterising statistics of
high-dimensional pdfs, as well as their divergences and distances,
where the pdf in the numerical implementation was assumed discretised on some regular grid.
We have demonstrated that high-dimensional pdfs,
pcfs, and some functions of them
can be approximated and represented in a low-rank tensor data format.
Utilisation of low-rank tensor techniques helps to reduce the computational complexity
and the storage cost from exponential $\C{O}(n^d)$ to linear in the dimension $d$, e.g.\
$O(d n r^2 )$ for the TT format. Here $n$ is the number of discretisation
points in one direction, $r<<n$ is the maximal tensor rank, and $d$ the problem dimension.
This document proposes a method for weakly supervised regression on uncertain datasets. It combines graph Laplacian regularization and cluster ensemble methodology. The method solves an auxiliary minimization problem to determine the optimal solution for predicting uncertain parameters. It is tested on artificial data to predict target values using a mixture of normal distributions with labeled, inaccurately labeled, and unlabeled samples. The method is shown to outperform a simplified version by reducing mean Wasserstein distance between predicted and true values.
Computing f-Divergences and Distances of High-Dimensional Probability Density...Alexander Litvinenko
Poster presented on Stochastic Numerics and Statistical Learning: Theory and Applications Workshop in KAUST, Saudi Arabia.
The task considered here was the numerical computation of characterising statistics of
high-dimensional pdfs, as well as their divergences and distances,
where the pdf in the numerical implementation was assumed discretised on some regular grid.
Even for moderate dimension $d$, the full storage and computation with such objects become very quickly infeasible.
We have demonstrated that high-dimensional pdfs,
pcfs, and some functions of them
can be approximated and represented in a low-rank tensor data format.
Utilisation of low-rank tensor techniques helps to reduce the computational complexity
and the storage cost from exponential $\C{O}(n^d)$ to linear in the dimension $d$, e.g.
O(d n r^2) for the TT format. Here $n$ is the number of discretisation
points in one direction, r<n is the maximal tensor rank, and d the problem dimension.
The particular data format is rather unimportant,
any of the well-known tensor formats (CP, Tucker, hierarchical Tucker, tensor-train (TT)) can be used,
and we used the TT data format. Much of the presentation and in fact the central train
of discussion and thought is actually independent of the actual representation.
In the beginning it was motivated through three possible ways how one may
arrive at such a representation of the pdf. One was if the pdf was given in some approximate
analytical form, e.g. like a function tensor product of lower-dimensional pdfs with a
product measure, or from an analogous representation of the pcf and subsequent use of the
Fourier transform, or from a low-rank functional representation of a high-dimensional
RV, again via its pcf.
The theoretical underpinnings of the relation between pdfs and pcfs as well as their
properties were recalled in Section: Theory, as they are important to be preserved in the
discrete approximation. This also introduced the concepts of the convolution and of
the point-wise multiplication Hadamard algebra, concepts which become especially important if
one wants to characterise sums of independent RVs or mixture models,
a topic we did not touch on for the sake of brevity but which follows very naturally from
the developments here. Especially the Hadamard algebra is also
important for the algorithms to compute various point-wise functions in the sparse formats.
Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...Alexander Litvinenko
Talk presented on SIAM IS 2022 conference.
Very often, in the course of uncertainty quantification tasks or
data analysis, one has to deal with high-dimensional random variables (RVs)
(with values in $\Rd$). Just like any other RV,
a high-dimensional RV can be described by its probability density (\pdf) and/or
by the corresponding probability characteristic functions (\pcf),
or a more general representation as
a function of other, known, random variables.
Here the interest is mainly to compute characterisations like the entropy, the Kullback-Leibler, or more general
$f$-divergences. These are all computed from the \pdf, which is often not available directly,
and it is a computational challenge to even represent it in a numerically
feasible fashion in case the dimension $d$ is even moderately large. It
is an even stronger numerical challenge to then actually compute said characterisations
in the high-dimensional case.
In this regard, in order to achieve a computationally feasible task, we propose
to approximate density by a low-rank tensor.
Low rank tensor approximation of probability density and characteristic funct...Alexander Litvinenko
This document summarizes a presentation on computing divergences and distances between high-dimensional probability density functions (pdfs) represented using tensor formats. It discusses:
1) Motivating the problem using examples from stochastic PDEs and functional representations of uncertainties.
2) Computing Kullback-Leibler divergence and other divergences when pdfs are not directly available.
3) Representing probability characteristic functions and approximating pdfs using tensor decompositions like CP and TT formats.
4) Numerical examples computing Kullback-Leibler divergence and Hellinger distance between Gaussian and alpha-stable distributions using these tensor approximations.
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
This document describes using the Continuation Multi-Level Monte Carlo (CMLMC) method to compute electromagnetic fields scattered from dielectric objects of uncertain shapes. CMLMC optimally balances statistical and discretization errors using fewer samples on fine meshes and more on coarse meshes. The method is tested by computing scattering cross sections for randomly perturbed spheres under plane wave excitation and comparing results to the unperturbed sphere. Computational costs and errors are analyzed to demonstrate the efficiency of CMLMC for this scattering problem with uncertain geometry.
1. Motivation: why do we need low-rank tensors
2. Tensors of the second order (matrices)
3. CP, Tucker and tensor train tensor formats
4. Many classical kernels have (or can be approximated in ) low-rank tensor format
5. Post processing: Computation of mean, variance, level sets, frequency
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
Computational tools for characterizing electromagnetic scattering from objects with uncertain shapes are needed in various applications ranging from remote sensing at microwave frequencies to Raman spectroscopy at optical frequencies. Often, such computational tools use the Monte Carlo (MC) method to sample a parametric space describing geometric uncertainties. For each sample, which corresponds to a realization of the geometry, a deterministic electromagnetic solver computes the scattered fields. However, for an accurate statistical characterization the number of MC samples has to be large. In this work, to address this challenge, the continuation multilevel Monte Carlo (\CMLMC) method is used together with a surface integral equation solver.
The \CMLMC method optimally balances statistical errors due to sampling of
the parametric space, and numerical errors due to the discretization of the geometry using a hierarchy of discretizations, from coarse to fine.
The number of realizations of finer discretizations can be kept low, with most samples
computed on coarser discretizations to minimize computational cost.
Consequently, the total execution time is significantly reduced, in comparison to the standard MC scheme.
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
Computational tools for characterizing electromagnetic scattering from objects with uncertain shapes are needed in various applications ranging from remote sensing at microwave frequencies to Raman spectroscopy at optical frequencies. Often, such computational tools use the Monte Carlo (MC) method to sample a parametric space describing geometric uncertainties. For each sample, which corresponds to a realization of the geometry, a deterministic electromagnetic solver computes the scattered fields. However, for an accurate statistical characterization the number of MC samples has to be large. In this work, to address this challenge, the continuation multilevel Monte Carlo (\CMLMC) method is used together with a surface integral equation solver.
The \CMLMC method optimally balances statistical errors due to sampling of
the parametric space, and numerical errors due to the discretization of the geometry using a hierarchy of discretizations, from coarse to fine.
The number of realizations of finer discretizations can be kept low, with most samples
computed on coarser discretizations to minimize computational cost.
Consequently, the total execution time is significantly reduced, in comparison to the standard MC scheme.
Propagation of Uncertainties in Density Driven Groundwater FlowAlexander Litvinenko
Major Goal: estimate risks of the pollution in a subsurface flow.
How?: we solve density-driven groundwater flow with uncertain porosity and permeability.
We set up density-driven groundwater flow problem,
review stochastic modeling and stochastic methods, use UG4 framework (https://gcsc.uni-frankfurt.de/simulation-and-modelling/ug4),
model uncertainty in porosity and permeability,
2D and 3D numerical experiments.
Simulation of propagation of uncertainties in density-driven groundwater flowAlexander Litvinenko
Consider stochastic modelling of the density-driven subsurface flow in 3D. This talk was presented by Dmitry Logashenko on the IMG conference in Kunming, China, August 2019.
This document summarizes a semi-supervised regression method that combines graph Laplacian regularization with cluster ensemble methodology. It proposes using a weighted averaged co-association matrix from the cluster ensemble as the similarity matrix in graph Laplacian regularization. The method (SSR-LRCM) finds a low-rank approximation of the co-association matrix to efficiently solve the regression problem. Experimental results on synthetic and real-world datasets show SSR-LRCM achieves significantly better prediction accuracy than an alternative method, while also having lower computational costs for large datasets. Future work will explore using a hierarchical matrix approximation instead of low-rank.
This document summarizes a talk on solving density-driven groundwater flow problems with uncertain porosity and permeability coefficients. The major goal is to estimate pollution risks in subsurface flows. The presentation covers: (1) setting up the groundwater flow problem; (2) reviewing stochastic modeling methods; (3) modeling uncertainty in porosity and permeability; (4) numerical methods to solve deterministic problems; and (5) 2D and 3D numerical experiments. The experiments demonstrate computing statistics of contaminant concentration and its propagation under uncertain parameters.
The 'Oedipus The King Student Revision Booklet' has been designed to help students prepare for writing about this text for a SAC or the exam. It scaffolds students to revise the plot, characters, symbols and dramatic devices of the text and builds their skills to write about the key ideas in response to a range of different types of essay topics.
CURRENT CASE COUNT: 880
• Texas: 729 (+5) (56% of cases are in Gaines County)
• New Mexico: 78 (+4) (83% of cases are from Lea County)
• Oklahoma: 17
• Kansas: 56 (38.89% of the cases are from Gray County)
HOSPITALIZATIONS: 103
• Texas: 94 - This accounts for 13% of all cases in the State.
• New Mexico: 7 – This accounts for 9.47% of all cases in New Mexico.
• Kansas: 2 - This accounts for 3.7% of all cases in Kansas.
DEATHS: 3
• Texas: 2 – This is 0.28% of all cases
• New Mexico: 1 – This is 1.35% of all cases
US NATIONAL CASE COUNT: 1,076 (confirmed and suspected)
INTERNATIONAL SPREAD
• Mexico: 1,753 (+198) 4 fatalities
‒ Chihuahua, Mexico: 1,657 (+167) cases, 3 fatalities, 9 hospitalizations
• Canada: 2518 (+239) (Includes Ontario’s outbreak, which began November 2024)
‒ Ontario, Canada: 1,795 (+173) 129 (+10) hospitalizations
‒ Alberta, Canada: 560 (+55)
Things to keep an eye on:
Mexico: Three children have died this month (all linked to the Chihuahua outbreak):
An 11-month-old and a 7-year-old with underlying conditions
A 1-year-old in Sonora whose family is from Chihuahua
Canada:
Ontario now reports more cases than the entire U.S.
Alberta’s case count continues to climb rapidly and is quickly closing in on 600 cases.
Emerging transmission chains in Manitoba and Saskatchewan underscore the need for vigilant monitoring of under-immunized communities and potential cross-provincial spread.
United States:
North Dakota: Grand Forks County has confirmed its first cases (2), linked to international travel. The state total is 21 since May 2 (including 4 in Cass County and 2 in Williams County), with one hospitalization reported.
OUTLOOK: With the spring–summer travel season peaking between Memorial Day and Labor Day, both domestic and international travel may fuel additional importations and spread. Although measles transmission is not strictly seasonal, crowded travel settings increase the risk for under-immunized individuals.
New syllabus entomology (Lession plan 121).pdfArshad Shaikh
*Fundamentals of Entomology*
Entomology is the scientific study of insects, including their behavior, ecology, evolution, classification, and management. Insects are the most diverse group of organisms on Earth, with over a million described species. Understanding entomology is crucial for managing insect pests, conserving beneficial insects, and appreciating their role in ecosystems.
*Key Concepts:*
- Insect morphology and anatomy
- Insect physiology and behavior
- Insect ecology and evolution
- Insect classification and identification
- Insect management and conservation
Entomology has numerous applications in agriculture, conservation, public health, and environmental science, making it a vital field of study.
How to Setup Renewal of Subscription in Odoo 18Celine George
A subscription is a recurring plan where you set a subscription period, such as weekly, monthly, or yearly. Based on this period, the subscription renews automatically. In Odoo 18, you have the flexibility to manage renewals either manually or automatically.
How to Add a Custom Menu, List view and FIlters in the Customer Portal Odoo 18Celine George
To create a new menu in the "My Accounts" portal and set up a list/form view in Odoo 18, follow these steps. For Example- The custom menu has now been added to the "Fleet" customer portal.
Flower Identification Class-10 by Kushal Lamichhane.pdfkushallamichhame
This includes the overall cultivation practices of rose prepared by:
Kushal Lamichhane
Instructor
Shree Gandhi Adarsha Secondary School
Kageshowri Manohara-09, Kathmandu, Nepal
This study describe how to write the Research Paper and its related issues. It also presents the major sections of Research Paper and various tools & techniques used for Polishing Research Paper
before final submission.
Finding a Right Journal and Publication Ethics are explain in brief.
Christian education is an important element in forming moral values, ethical Behaviour and
promoting social unity, especially in diverse nations like in the Caribbean. This study examined
the impact of Christian education on the moral growth in the Caribbean, characterized by
significant Christian denomination, like the Orthodox, Catholic, Methodist, Lutheran and
Pentecostal. Acknowledging the historical and social intricacies in the Caribbean, this study
tends to understand the way in which Christian education mold ethical decision making, influence interpersonal relationships and promote communal values. These studies’ uses, qualitative and quantitative research method to conduct semi-structured interviews for twenty
(25) Church respondents which cut across different age groups and genders in the Caribbean. A
thematic analysis was utilized to identify recurring themes related to ethical Behaviour, communal values and moral development. The study analyses the three objectives of the study:
how Christian education Mold’s ethical Behaviour and enhance communal values, the role of
Christian educating in promoting ecumenism and the effect of Christian education on moral
development. Moreover, the findings show that Christian education serves as a fundamental role
for personal moral evaluation, instilling a well-structured moral value, promoting good
Behaviour and communal responsibility such as integrity, compassion, love and respect. However, the study also highlighted challenges including biases in Christian teachings, exclusivity and misconceptions about certain practices, which impede the actualization of
How to create and manage blogs in odoo 18Celine George
A blog serves as a space for sharing articles and information.
In Odoo 18, users can easily create and publish blogs through
the blog menu. This guide offers step-by-step instructions on
setting up and managing a blog on an Odoo 18 website.
Jack Lutkus is an education champion, community-minded innovator, and cultural enthusiast. A social work graduate student at Aurora University, he also holds a BA from the University of Iowa.
Paper 110A | Shadows and Light: Exploring Expressionism in ‘The Cabinet of Dr...Rajdeep Bavaliya
Dive into the haunting worlds of German Expressionism as we unravel how shadows and light elevate ‘The Cabinet of Dr. Caligari’ and ‘Nosferatu: A Symphony of Horror’ into timeless masterpieces. Discover the psychological power of chiaroscuro, distorted sets, and evocative silhouettes that shaped modern horror. Whether you’re a film buff or a budding cinephile, this journey through post‑WWI trauma and surreal visuals will leave you seeing movies in a whole new light. Hit play, share your favorite shock‑and‑awe moment in the comments, and don’t forget to follow for more deep‑dives into cinema’s most influential movements!
M.A. Sem - 2 | Presentation
Presentation Season - 2
Paper - 110A: History of English Literature – From 1900 to 2000
Submitted Date: April 1, 2025
Paper Name: History of English Literature – From 1900 to 2000
Topic: Shadows and Light: Exploring Expressionism in ‘The Cabinet of Dr. Caligari’ and ‘Nosferatu: A Symphony of Horror’
[Please copy the link and paste it into any web browser to access the content.]
Video Link: https://youtu.be/pWjHqo6clT4
For a more in-depth discussion of this presentation, please visit the full blog post at the following link:
Please visit this blog to explore additional presentations from this season:
Hashtags:
#GermanExpressionism #SilentHorror #Caligari #Nosferatu #Chiaroscuro #VisualStorytelling #FilmHistory #HorrorCinema #CinematicArt #ExpressionistAesthetics
Keyword Tags:
Expressionism, The Cabinet of Dr. Caligari, Nosferatu, silent film horror, film noir origins, German Expressionist cinema, chiaroscuro techniques, cinematic shadows, psychological horror, visual aesthetics
The PDF titled "Critical Thinking and Bias" by Jibi Moses aims to equip a diverse audience from South Sudan with the knowledge and skills necessary to identify and challenge biases and stereotypes. It focuses on developing critical thinking abilities and promoting inclusive attitudes to foster a more cohesive and just society. It defines bias as a tendency or prejudice affecting perception and interactions, categorizing it into conscious and unconscious (implicit) biases. The content highlights the impact of societal and cultural conditioning on these biases, particularly within the South Sudanese context.
TechSoup Microsoft Copilot Nonprofit Use Cases and Live Demo - 2025.05.28.pdfTechSoup
In this webinar we will dive into the essentials of generative AI, address key AI concerns, and demonstrate how nonprofits can benefit from using Microsoft’s AI assistant, Copilot, to achieve their goals.
This event series to help nonprofits obtain Copilot skills is made possible by generous support from Microsoft.
ISO 27001 Lead Auditor Exam Practice Questions and Answers-.pdfinfosec train
🧠 𝐏𝐫𝐞𝐩𝐚𝐫𝐢𝐧𝐠 𝐟𝐨𝐫 𝐭𝐡𝐞 𝐈𝐒𝐎 𝟐𝟕𝟎𝟎𝟏 𝐋𝐞𝐚𝐝 𝐀𝐮𝐝𝐢𝐭𝐨𝐫 𝐄𝐱𝐚𝐦? 𝐃𝐨𝐧’𝐭 𝐉𝐮𝐬𝐭 𝐒𝐭𝐮𝐝𝐲—𝐏𝐫𝐚𝐜𝐭𝐢𝐜𝐞 𝐰𝐢𝐭𝐡 𝐏𝐮𝐫𝐩𝐨𝐬𝐞!
We’ve compiled a 𝐜𝐨𝐦𝐩𝐫𝐞𝐡𝐞𝐧𝐬𝐢𝐯𝐞 𝐰𝐡𝐢𝐭𝐞 𝐩𝐚𝐩𝐞𝐫 featuring 𝐫𝐞𝐚𝐥𝐢𝐬𝐭𝐢𝐜, 𝐬𝐜𝐞𝐧𝐚𝐫𝐢𝐨-𝐛𝐚𝐬𝐞𝐝 𝐩𝐫𝐚𝐜𝐭𝐢𝐜𝐞 𝐪𝐮𝐞𝐬𝐭𝐢𝐨𝐧𝐬 𝐚𝐧𝐝 𝐚𝐧𝐬𝐰𝐞𝐫𝐬 designed specifically for those targeting the 𝐈𝐒𝐎/𝐈𝐄𝐂 𝟐𝟕𝟎𝟎𝟏 𝐋𝐞𝐚𝐝 𝐀𝐮𝐝𝐢𝐭𝐨𝐫 𝐜𝐞𝐫𝐭𝐢𝐟𝐢𝐜𝐚𝐭𝐢𝐨𝐧.
🔍 𝐈𝐧𝐬𝐢𝐝𝐞 𝐲𝐨𝐮'𝐥𝐥 𝐟𝐢𝐧𝐝:
✅ Exam-style questions mapped to ISO 27001:2022
✅ Detailed explanations (not just the right answer—but why it’s right)
✅ Mnemonics, control references (like A.8.8, A.5.12, A.8.24), and study tips
✅ Key audit scenarios: nonconformities, SoA vs scope, AART treatment options, CIA triad, and more
𝐖𝐡𝐞𝐭𝐡𝐞𝐫 𝐲𝐨𝐮'𝐫𝐞:
🔹 Starting your ISO journey
🔹 Preparing for your Lead Auditor exam
🔹 Or mentoring others in information security audits...
This guide can seriously boost your confidence and performance.
Education Funding Equity in North Carolina: Looking Beyond IncomeEducationNC
Application of parallel hierarchical matrices for parameter inference and prediction
1. Parallel Hierarchical Matrix Technique to
Approximate Large Covariance Matrices, Likelihood
Functions and Parameter Identification
Alexander Litvinenko1,
(joint work with V. Berikov2, M. Genton3, D. Keyes3, R.
Kriemann4, and Y. Sun3)
SIAM CSE 2021
1RWTH Aachen, Germany, 2Novosibirsk Staate University, 3KAUST, Saudi Arabia,
4MPI for Mathematics in the Sciences in Leipzig, Germany
3. 4
*
The structure of the talk
1. Motivation: improve statistical models, data analysis,
prediction
2. Identification of unknown parameters via maximizing Gaussian
log-likelihood (MLE)
3. Tools: Hierarchical matrices [Hackbusch 1999]
4. Matérn covariance function, joint Gaussian log-likelihood
5. Error analysis
6. Prediction at new locations
7. Comparison with machine learning methods
3
4. 4
*
Identifying unknown parameters
Given:
Let s1, . . . , sn be locations.
Z = {Z(s1), . . . , Z(sn)}>, where Z(s) is a Gaussian random field
indexed by a spatial location s ∈ Rd , d ≥ 1.
Assumption: Z has mean zero and stationary parametric
covariance function, e.g. Matérn:
C(θ) =
2σ2
Γ(ν)
r
2`
ν
Kν
r
`
+ τ2
I, θ = (σ2
, ν, `, τ2
).
To identify: unknown parameters θ := (σ2, ν, `, τ2).
4
5. 4
*
Identifying unknown parameters
Statistical inference about θ is then based on the Gaussian
log-likelihood function:
L(C(θ)) = L(θ) = −
n
2
log(2π) −
1
2
log|C(θ)| −
1
2
Z
C(θ)−1
Z, (1)
where the covariance matrix C(θ) has entries
C(si − sj ; θ), i, j = 1, . . . , n.
The maximum likelihood estimator of θ is the value b
θ that
maximizes (1).
5
7. 4
*
What will change after H-matrix approximation?
Approximate C by CH
1. How the eigenvalues of C and CH differ ?
2. How det(C) differs from det(CH) ?
3. How L differs from LH ? [Mario Bebendorf et al]
4. How C−1 differs from (CH)−1 ? [Mario Bebendorf et al]
5. How L̃(θ, k) differs from L(θ)?
6. What is optimal H-matrix rank?
7. How θH differs from θ?
For theory, estimates for the rank and accuracy see works of
Bebendorf, Grasedyck, Le Borne, Hackbusch,...
7
8. 4
*
Details of the parameter identification
To maximize the log-likelihood function we use the Brent’s method
(combining bisection method, secant method and inverse quadratic
interpolation) or any other.
1. C(θ) ≈ e
C(θ, ε).
2. e
C(θ, k) = e
L(θ, k)e
L(θ, k)T
3. ZT e
C−1Z = ZT (e
Le
LT )−1Z = vT · v, where v is a solution of
e
L(θ, k)v(θ) := Z.
log det{e
C} = log det{e
Le
LT
} = log det{
n
Y
i=1
λ2
i } = 2
n
X
i=1
logλi ,
e
L(θ, k) = −
n
2
log(2π) −
n
X
i=1
log{e
Lii (θ, k)} −
1
2
(v(θ)T
· v(θ)). (2)
8
9. Dependence of log-likelihood ingredients on parameters, n = 4225.
k = 8 in the first row and k = 16 in the second.
9
10. 4
*
Remark: stabilization with nugget
To avoid instability in computing Cholesky, we add: e
Cm = e
C + τ2I.
Let λi be eigenvalues of e
C, then eigenvalues of e
Cm will be λi + τ2,
log det(e
Cm) = log
Qn
i=1(λi + τ2) =
Pn
i=1 log(λi + τ2).
(left) Dependence of the negative log-likelihood on parameter `
with nuggets {0.01, 0.005, 0.001} for the Gaussian covariance.
(right) Zoom of the left figure near minimum; n = 2000 random
points from moisture example, rank k = 14, τ2 = 1, ν = 0.5.
10
11. 4
*
Error analysis
Theorem (1)
Let e
C be an H-matrix approximation of matrix C ∈ Rn×n such that
ρ(e
C−1
C − I) ≤ ε 1.
Then
|log det C − log det e
C| ≤ −nlog(1 − ε), (3)
Proof: See [Ballani, Kressner 14] and [Ipsen 05].
Remark: factor n is pessimistic and is not really observed
numerically.
11
12. 4
*
Error in the log-likelihood
Theorem (2)
Let e
C ≈ C ∈ Rn×n and Z be a vector, kZk ≤ c0 and kC−1k ≤ c1.
Let ρ(e
C−1C − I) ≤ ε 1. Then it holds
| e
L(θ) − L(θ)| =
1
2
(log|C| − log|e
C|) +
1
2
|ZT
C−1
− e
C−1
Z|
≤ −
1
2
· nlog(1 − ε) +
1
2
|ZT
C−1
C − e
C−1
C
C−1
Z|
≤ −
1
2
· nlog(1 − ε) +
1
2
c2
0 · c1 · ε.
12
13. 4
*
H-matrix approximation
ε accuracy in each sub-block, n = 16641, ν = 0.5,
c.r.=compression ratio.
ε |log|C| − log|e
C|| |
log|C|−log|e
C|
log|e
C|
| kC − e
CkF
kC−e
Ck2
kCk2
kI − (e
Le
L
)−1
Ck2 c.r. in %
` = 0.0334
1e-1 3.2e-4 1.2e-4 7.0e-3 7.6e-3 2.9 9.16
1e-2 1.6e-6 6.0e-7 1.0e-3 6.7e-4 9.9e-2 9.4
1e-4 1.8e-9 7.0e-10 1.0e-5 7.3e-6 2.0e-3 10.2
1e-8 4.7e-13 1.8e-13 1.3e-9 6e-10 2.1e-7 12.7
` = 0.2337
1e-4 9.8e-5 1.5e-5 8.1e-5 1.4e-5 2.5e-1 9.5
1e-8 1.45e-9 2.3e-10 1.1e-8 1.5e-9 4e-5 11.3
log|C| = 2.63 for ` = 0.0334 and log|C| = 6.36 for ` = 0.2337.
13
14. 4
*
Boxplots for unknown parameters
Moisture data example. Boxplots for the 100 estimates of (`, ν,
σ2), respectively, when n = 32K, 16K, 8K, 4K, 2K. H-matrix with
a fixed rank k = 11. Green horizontal lines denote 25% and 75%
quantiles for n = 32K.
14
15. 4
*
How much memory is needed?
0 0.5 1 1.5 2 2.5
ℓ, ν = 0.325, σ2
= 0.98
5
5.5
6
6.5
size,
in
bytes
×10 6
1e-4
1e-6
0.2 0.4 0.6 0.8 1 1.2 1.4
ν, ℓ = 0.58, σ2
= 0.98
5.2
5.4
5.6
5.8
6
6.2
6.4
6.6
size,
in
bytes
×10 6
1e-4
1e-6
(left) Dependence of the matrix size on the covariance length `,
and (right) the smoothness ν for two different H-accuracies
ε = {10−4, 10−6}
15
17. 4
*
How log-likelihood depends on n?
ν
0 0.1 0.2 0.3 0.4 0.5 0.6
−L
10 3
10 4
10 5
2000
4000
8000
16000
32000
Figure: Dependence of negative log-likelihood function on different
number of locations n = {2000, 4000, 8000, 16000, 32000} in log-scale.
17
18. 4
*
Time and memory for parallel H-matrix approximation
Maximal # cores is 40, ν = 0.325, ` = 0.64, σ2 = 0.98
n C̃ L̃L̃
time size kB/dof time size kI − (L̃L̃
)−1
C̃k2
sec. MB sec. MB
32.000 3.3 162 5.1 2.4 172.7 2.4 · 10−3
128.000 13.3 776 6.1 13.9 881.2 1.1 · 10−2
512.000 52.8 3420 6.7 77.6 4150 3.5 · 10−2
2.000.000 229 14790 7.4 473 18970 1.4 · 10−1
Dell Station, 20 × 2 cores, 128 GB RAM, bought in 2013 for
10.000 USD.
18
19. 4
*
Prediction
Let Z = (Z1, Z2) has mean zero and a stationary covariance, Z1 -
known, Z2 unknown.
C =
C11 C12
C21 C22,
We compute predicted values
Z2 = C21C−1
11 Z1
Z2 has the conditional distribution with the mean value C21C−1
11 Z1
and the covariance matrix C22 − C21C−1
11 C12.
19
20. 4
*
2021 KAUST Competition
We participated in 2021 KAUST Competition on Spatial
Statistics for Large Datasets.
You can download the datasets and look the final results here
https://cemse.kaust.edu.sa/stsds/
2021-kaust-competition-spatial-statistics-large-datasets
20
21. 4
*
Prediction
Prediction for two datasets. The yellow points at 900.000 locations
were used for training and the blue points were predicted at
100.000 new locations.
21
22. 4
*
Machine learning methods to make prediction
k-nearest neighbours (kNN): For each point x find its k nearest
neighbors x1, . . . , xk, and set: ŷ(x) = 1
k
Pk
i=1 yi .
Random Forest (RF): a large number of decision (or regression)
trees are generated independently on random sub-samples of data.
The final decision for x is calculated over the ensemble of trees by
averaging the predicted outcomes.
Deep Neural Network (DNN): fully connected neural network
Input layer consists of two neurons (input feature dimensionality),
and output layer consists of one neuron (predicted feature
dimensionality).
22
23. 4
*
Prediction by kNN
Prediction obtained by the kNN method. The yellow points at
900.000 locations were used for training and the blue points were
predicted at 100.000 new locations. One can see a very well
alignment of both.
23
24. 4
*
Conclusion
I With H-matrices you can approximate Matérn covariance
matrices, Gaussian log-likelihoods, identify unknown
parameters and make predictions
I MLE estimate and predictions depend on H-matrix accuracy
I parameter identification problem has multiple solutions
I Investigated dependence of H-matrix approximation error on
the estimated parameters
I Each of ML methods needs fine-tuning stage to optimize its
hyperparameters or architecture.
24
25. 4
*
Open questions and TODOs
I The Gaussian log-likelihood function has some drawbacks for
very large matrices
I How to skip/avoid redundant data?
I A good starting point for optimization is needed
I a “preconditioner” (a simple cov. matrix) is needed
I H-matrices become expensive for large number of parameters
to be identified
I error estimates are needed
26. 4
*
Literature
All tests are reproducible
https://github.com/litvinen/large_random_fields.git
1. A. Litvinenko, R. Kriemann, M.G. Genton, Y. Sun, D.E. Keyes, HLIBCov:
Parallel hierarchical matrix approximation of large covariance matrices
and likelihoods with applications in parameter identification, MethodsX 7,
100600, 2020
2. A. Litvinenko, Y. Sun, M.G. Genton, D.E. Keyes, Likelihood
approximation with hierarchical matrices for large spatial datasets,
Computational Statistics Data Analysis 137, 115-132, 2019
3. A. Litvinenko, R. Kriemann, V. Berikov, Identification of unknown
parameters and prediction with hierarchical matrices, look arXiv and
ResearchGate, March 2021
4. M. Espig, W. Hackbusch, A. Litvinenko, H.G. Matthies, E. Zander,
Iterative algorithms for the post-processing of high-dimensional data,
Journal of Computational Physics 410, 109396, 2020
26
27. 4
*
Literature
5. A. Litvinenko, D. Keyes, V. Khoromskaia, B.N. Khoromskij, H.G.
Matthies, Tucker tensor analysis of Matérn functions in spatial statistics,
Computational Methods in Applied Mathematics 19 (1), 101-122, 2019
6. B. N. Khoromskij, A. Litvinenko, H.G. Matthies, Application of
hierarchical matrices for computing the Karhunen-Loéve expansion,
Computing 84 (1-2), 49-67, 31, 2009
7. W. Nowak, A. Litvinenko, Kriging and spatial design accelerated by
orders of magnitude: Combining low-rank covariance approximations with
FFT-techniques, Mathematical Geosciences 45 (4), 411-435, 2013
8. M. Espig, W. Hackbusch, A. Litvinenko, H.G. Matthies, E. Zander,
Efficient analysis of high dimensional data in tensor formats, Sparse Grids
and Applications, 31-56, Springer, Berlin, 2013
27
28. 4
*
Acknowledgement
V. Berikov was supported by the state contract of the Sobolev
Institute of Mathematics (project no 0314-2019-0015), and by
RFBR grant 19-29-01175.
A. Litvinenko was supported by funding from the Alexander von
Humboldt Foundation.
28