1 |
Three-dimensional Modeling and Simulation of a Tuning ForkLarisch, Lukas 16 September 2018 (has links)
The mathematical characterization of the sound of a musical instrument still follows Schumann’s laws [1]. According to this theory, the resonances of the instrument body, “the formants”, filter the oscillations of the sound generator (e.g., strings) and produce the characteristic “timbre” of an instrument. This is a strong simplification of the actual situation. It applies to a point source and does not distinguish between a loudspeaker and a three-dimensional instrument.
In this work we investigate Finite-Element-based numerical simulations of eigenfrequencies and eigenmodes of a tuning fork in order to capture the oscillation behavior of its eigenfrequencies. We model the tuning fork as an elastic solid body and solve an eigenvalue equation derived from a system of coupled equations from linear elasticity theory on an unstructured three-dimensional grid. The eigenvalue problem is solved using the preconditioned inverse iteration (PINVIT) method with an efficient geometric multigrid (GMG) preconditioner. The latter allows us to resolve the tuning fork with a high resolution grid, which is required to capture fine modes of the simulated eigenfrequencies. To verify our results, we compare them with measurement data obtained from an experimental modal analyses of a real reference tuning fork.
It turns out that our model is sufficient to capture the first eight eigenmodes of a reference tuning fork, whose identification and reproduction by simulation is novel to the knowledge of the author.
|
2 |
Eigenvalue Algorithms for Symmetric Hierarchical Matrices / Eigenwert-Algorithmen für Symmetrische Hierarchische MatrizenMach, Thomas 05 April 2012 (has links) (PDF)
This thesis is on the numerical computation of eigenvalues of symmetric hierarchical matrices. The numerical algorithms used for this computation are derivations of the LR Cholesky algorithm, the preconditioned inverse iteration, and a bisection method based on LDLT factorizations.
The investigation of QR decompositions for H-matrices leads to a new QR decomposition. It has some properties that are superior to the existing ones, which is shown by experiments using the HQR decompositions to build a QR (eigenvalue) algorithm for H-matrices does not progress to a more efficient algorithm than the LR Cholesky algorithm.
The implementation of the LR Cholesky algorithm for hierarchical matrices together with deflation and shift strategies yields an algorithm that require O(n) iterations to find all eigenvalues. Unfortunately, the local ranks of the iterates show a strong growth in the first steps. These H-fill-ins makes the computation expensive, so that O(n³) flops and O(n²) storage are required.
Theorem 4.3.1 explains this behavior and shows that the LR Cholesky algorithm is efficient for the simple structured Hl-matrices.
There is an exact LDLT factorization for Hl-matrices and an approximate LDLT factorization for H-matrices in linear-polylogarithmic complexity. This factorizations can be used to compute the inertia of an H-matrix. With the knowledge of the inertia for arbitrary shifts, one can compute an eigenvalue by bisectioning. The slicing the spectrum algorithm can compute all eigenvalues of an Hl-matrix in linear-polylogarithmic complexity. A single eigenvalue can be computed in O(k²n log^4 n).
Since the LDLT factorization for general H-matrices is only approximative, the accuracy of the LDLT slicing algorithm is limited. The local ranks of the LDLT factorization for indefinite matrices are generally unknown, so that there is no statement on the complexity of the algorithm besides the numerical results in Table 5.7.
The preconditioned inverse iteration computes the smallest eigenvalue and the corresponding eigenvector. This method is efficient, since the number of iterations is independent of the matrix dimension.
If other eigenvalues than the smallest are searched, then preconditioned inverse iteration can not be simply applied to the shifted matrix, since positive definiteness is necessary. The squared and shifted matrix (M-mu I)² is positive definite. Inner eigenvalues can be computed by the combination of folded spectrum method and PINVIT. Numerical experiments show that the approximate inversion of (M-mu I)² is more expensive than the approximate inversion of M, so that the computation of the inner eigenvalues is more expensive.
We compare the different eigenvalue algorithms. The preconditioned inverse iteration for hierarchical matrices is better than the LDLT slicing algorithm for the computation of the smallest eigenvalues, especially if the inverse is already available. The computation of inner eigenvalues with the folded spectrum method and preconditioned inverse iteration is more expensive. The LDLT slicing algorithm is competitive to H-PINVIT for the computation of inner eigenvalues.
In the case of large, sparse matrices, specially tailored algorithms for sparse matrices, like the MATLAB function eigs, are more efficient.
If one wants to compute all eigenvalues, then the LDLT slicing algorithm seems to be better than the LR Cholesky algorithm. If the matrix is small enough to be handled in dense arithmetic (and is not an Hl(1)-matrix), then dense eigensolvers, like the LAPACK function dsyev, are superior.
The H-PINVIT and the LDLT slicing algorithm require only an almost linear amount of storage. They can handle larger matrices than eigenvalue algorithms for dense matrices.
For Hl-matrices of local rank 1, the LDLT slicing algorithm and the LR Cholesky algorithm need almost the same time for the computation of all eigenvalues. For large matrices, both algorithms are faster than the dense LAPACK function dsyev.
|
3 |
Eigenvalue Algorithms for Symmetric Hierarchical MatricesMach, Thomas 20 February 2012 (has links)
This thesis is on the numerical computation of eigenvalues of symmetric hierarchical matrices. The numerical algorithms used for this computation are derivations of the LR Cholesky algorithm, the preconditioned inverse iteration, and a bisection method based on LDLT factorizations.
The investigation of QR decompositions for H-matrices leads to a new QR decomposition. It has some properties that are superior to the existing ones, which is shown by experiments using the HQR decompositions to build a QR (eigenvalue) algorithm for H-matrices does not progress to a more efficient algorithm than the LR Cholesky algorithm.
The implementation of the LR Cholesky algorithm for hierarchical matrices together with deflation and shift strategies yields an algorithm that require O(n) iterations to find all eigenvalues. Unfortunately, the local ranks of the iterates show a strong growth in the first steps. These H-fill-ins makes the computation expensive, so that O(n³) flops and O(n²) storage are required.
Theorem 4.3.1 explains this behavior and shows that the LR Cholesky algorithm is efficient for the simple structured Hl-matrices.
There is an exact LDLT factorization for Hl-matrices and an approximate LDLT factorization for H-matrices in linear-polylogarithmic complexity. This factorizations can be used to compute the inertia of an H-matrix. With the knowledge of the inertia for arbitrary shifts, one can compute an eigenvalue by bisectioning. The slicing the spectrum algorithm can compute all eigenvalues of an Hl-matrix in linear-polylogarithmic complexity. A single eigenvalue can be computed in O(k²n log^4 n).
Since the LDLT factorization for general H-matrices is only approximative, the accuracy of the LDLT slicing algorithm is limited. The local ranks of the LDLT factorization for indefinite matrices are generally unknown, so that there is no statement on the complexity of the algorithm besides the numerical results in Table 5.7.
The preconditioned inverse iteration computes the smallest eigenvalue and the corresponding eigenvector. This method is efficient, since the number of iterations is independent of the matrix dimension.
If other eigenvalues than the smallest are searched, then preconditioned inverse iteration can not be simply applied to the shifted matrix, since positive definiteness is necessary. The squared and shifted matrix (M-mu I)² is positive definite. Inner eigenvalues can be computed by the combination of folded spectrum method and PINVIT. Numerical experiments show that the approximate inversion of (M-mu I)² is more expensive than the approximate inversion of M, so that the computation of the inner eigenvalues is more expensive.
We compare the different eigenvalue algorithms. The preconditioned inverse iteration for hierarchical matrices is better than the LDLT slicing algorithm for the computation of the smallest eigenvalues, especially if the inverse is already available. The computation of inner eigenvalues with the folded spectrum method and preconditioned inverse iteration is more expensive. The LDLT slicing algorithm is competitive to H-PINVIT for the computation of inner eigenvalues.
In the case of large, sparse matrices, specially tailored algorithms for sparse matrices, like the MATLAB function eigs, are more efficient.
If one wants to compute all eigenvalues, then the LDLT slicing algorithm seems to be better than the LR Cholesky algorithm. If the matrix is small enough to be handled in dense arithmetic (and is not an Hl(1)-matrix), then dense eigensolvers, like the LAPACK function dsyev, are superior.
The H-PINVIT and the LDLT slicing algorithm require only an almost linear amount of storage. They can handle larger matrices than eigenvalue algorithms for dense matrices.
For Hl-matrices of local rank 1, the LDLT slicing algorithm and the LR Cholesky algorithm need almost the same time for the computation of all eigenvalues. For large matrices, both algorithms are faster than the dense LAPACK function dsyev.:List of Figures xi
List of Tables xiii
List of Algorithms xv
List of Acronyms xvii
List of Symbols xix
Publications xxi
1 Introduction 1
1.1 Notation 2
1.2 Structure of this Thesis 3
2 Basics 5
2.1 Linear Algebra and Eigenvalues 6
2.1.1 The Eigenvalue Problem 7
2.1.2 Dense Matrix Algorithms 9
2.2 Integral Operators and Integral Equations 14
2.2.1 Definitions 14
2.2.2 Example - BEM 16
2.3 Introduction to Hierarchical Arithmetic 17
2.3.1 Main Idea 17
2.3.2 Definitions 19
2.3.3 Hierarchical Arithmetic 24
2.3.4 Simple Hierarchical Matrices (Hl-Matrices) 30
2.4 Examples 33
2.4.1 FEM Example 33
2.4.2 BEM Example 36
2.4.3 Randomly Generated Examples 37
2.4.4 Application Based Examples 38
2.4.5 One-Dimensional Integral Equation 38
2.5 Related Matrix Formats 39
2.5.1 H2-Matrices 40
2.5.2 Diagonal plus Semiseparable Matrices 40
2.5.3 Hierarchically Semiseparable Matrices 42
2.6 Review of Existing Eigenvalue Algorithms 44
2.6.1 Projection Method 44
2.6.2 Divide-and-Conquer for Hl(1)-Matrices 45
2.6.3 Transforming Hierarchical into Semiseparable Matrices 46
2.7 Compute Cluster Otto 47
3 QR Decomposition of Hierarchical Matrices 49
3.1 Introduction 49
3.2 Review of Known QR Decompositions for H-Matrices 50
3.2.1 Lintner’s H-QR Decomposition 50
3.2.2 Bebendorf’s H-QR Decomposition 52
3.3 A new Method for Computing the H-QR Decomposition 54
3.3.1 Leaf Block-Column 54
3.3.2 Non-Leaf Block Column 56
3.3.3 Complexity 57
3.3.4 Orthogonality 60
3.3.5 Comparison to QR Decompositions for Sparse Matrices 61
3.4 Numerical Results 62
3.4.1 Lintner’s H-QR decomposition 62
3.4.2 Bebendorf’s H-QR decomposition 66
3.4.3 The new H-QR decomposition 66
3.5 Conclusions 67
4 QR-like Algorithms for Hierarchical Matrices 69
4.1 Introduction 70
4.1.1 LR Cholesky Algorithm 70
4.1.2 QR Algorithm 70
4.1.3 Complexity 71
4.2 LR Cholesky Algorithm for Hierarchical Matrices 72
4.2.1 Algorithm 72
4.2.2 Shift Strategy 72
4.2.3 Deflation 73
4.2.4 Numerical Results 73
4.3 LR Cholesky Algorithm for Diagonal plus Semiseparable Matrices 75
4.3.1 Theorem 75
4.3.2 Application to Tridiagonal and Band Matrices 79
4.3.3 Application to Matrices with Rank Structure 79
4.3.4 Application to H-Matrices 80
4.3.5 Application to Hl-Matrices 82
4.3.6 Application to H2-Matrices 83
4.4 Numerical Examples 84
4.5 The Unsymmetric Case 84
4.6 Conclusions 88
5 Slicing the Spectrum of Hierarchical Matrices 89
5.1 Introduction 89
5.2 Slicing the Spectrum by LDLT Factorization 91
5.2.1 The Function nu(M − µI) 91
5.2.2 LDLT Factorization of Hl-Matrices 92
5.2.3 Start-Interval [a, b] 96
5.2.4 Complexity 96
5.3 Numerical Results 97
5.4 Possible Extensions 100
5.4.1 LDLT Slicing Algorithm for HSS Matrices 103
5.4.2 LDLT Slicing Algorithm for H-Matrices 103
5.4.3 Parallelization 105
5.4.4 Eigenvectors 107
5.5 Conclusions 107
6 Computing Eigenvalues by Vector Iterations 109
6.1 Power Iteration 109
6.1.1 Power Iteration for Hierarchical Matrices 110
6.1.2 Inverse Iteration 111
6.2 Preconditioned Inverse Iteration for Hierarchical Matrices 111
6.2.1 Preconditioned Inverse Iteration 113
6.2.2 The Approximate Inverse of an H-Matrix 115
6.2.3 The Approximate Cholesky Decomposition of an H-Matrix 116
6.2.4 PINVIT for H-Matrices 117
6.2.5 The Interior of the Spectrum 120
6.2.6 Numerical Results 123
6.2.7 Conclusions 130
7 Comparison of the Algorithms and Numerical Results 133
7.1 Theoretical Comparison 133
7.2 Numerical Comparison 135
8 Conclusions 141
Theses 143
Bibliography 145
Index 153
|
Page generated in 0.0171 seconds