• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 560
  • 32
  • 6
  • 2
  • Tagged with
  • 620
  • 620
  • 585
  • 52
  • 41
  • 40
  • 38
  • 34
  • 33
  • 30
  • 30
  • 29
  • 28
  • 27
  • 26
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
301

Quasi-Arithmetic Filters for Topology Optimization / Kvasiaritmetiska filter för topologioptimering

Hägg, Linus January 2016 (has links)
Topology optimization is a framework for finding the optimal layout of material within a given region of space. In material distribution topology optimization, a material indicator function determines the material state at each point within the design domain. It is well known that naive formulations of continuous material distribution topology optimization problems often lack solutions. To obtain numerical solutions, the continuous problem is approximated by a finite-dimensional problem. The finite-dimensional approximation is typically obtained by partitioning the design domain into a finite number of elements and assigning to each element a design variable that determines the material state of that element. Although the finite-dimensional problem generally is solvable, a sequence of solutions corresponding to ever finer partitions of the design domain may not converge; that is, the optimized designs may exhibit mesh-dependence. Filtering procedures are amongst the most popular methods used to handle the existence issue related to the continuous problem as well as the mesh-dependence related to the finite-dimensional approximation. Over the years, a variety of filters for topology optimization have been presented. To harmonize the use and analysis of filters within the field of topology optimization, we introduce the class of fW-mean filters that is based on the weighted quasi-arithmetic mean, also known as the weighted generalized f-mean, over some neighborhoods. We also define the class of generalized fW-mean filters that contains the vast majority of filters for topology optimization. In particular, the class of generalized fW-mean filters includes the fW-mean filters, as well as the projected fW-mean filters that are formed by adding a projection step to the fW-mean filters. If the design variables are located in a regular grid, uniform weights are used within each neighborhood, and equal sized polytope shaped neighborhoods are used, then a cascade of generalized fW-mean filters can be applied with a computational complexity that is linear in the number of design variables. Detailed algorithms for octagonal shaped neighborhoods in 2D and rhombicuboctahedron shaped neighborhoods in 3D are provided. The theoretically obtained computational complexity of the algorithm for octagonal shaped neighborhoods in 2D has been numerically verified. By using the same type of algorithm as for filtering, the additional computational complexity for computing derivatives needed in gradient based optimization is also linear in the number of design variables. To exemplify the use of generalized fW-mean filters in topology optimization, we consider minimization of compliance (maximization of global stiffness) of linearly elastic continuum bodies. We establish the existence of solutions to a version of the continuous minimal compliance problem when a cascade of projected continuous fW-mean filters is included in the formulation. Bourdin's classical existence result for the linear density filter is a partial case of this general theorem for projected continuous fW-mean filters. Inspired by the works of Svanberg & Svärd and Sigmund, we introduce the harmonic open-close filter, which is a cascade of four fW-mean filters. We present large-scale numerical experiments indicating that, for minimal compliance problems, the harmonic open-close filter produces almost binary designs, provides independent size control on both material and void regions, and yields mesh-independent designs.
302

Skew-symmetric matrix pencils : stratification theory and tools

Dmytryshyn, Andrii January 2014 (has links)
Investigating the properties, explaining, and predicting the behaviour of a physical system described by a system (matrix) pencil often require the understanding of how canonical structure information of the system pencil may change, e.g., how eigenvalues coalesce or split apart, due to perturbations in the matrix pencil elements. Often these system pencils have different block-partitioning and / or symmetries. We study changes of the congruence canonical form of a complex skew-symmetric matrix pencil under small perturbations. The problem of computing the congruence canonical form is known to be ill-posed: both the canonical form and the reduction transformation depend discontinuously on the entries of a pencil. Thus it is important to know the canonical forms of all such pencils that are close to the investigated pencil. One way to investigate this problem is to construct the stratification of orbits and bundles of the pencils. To be precise, for any problem dimension we construct the closure hierarchy graph for congruence orbits or bundles. Each node (vertex) of the graph represents an orbit (or a bundle) and each edge represents the cover/closure relation. Such a relation means that there is a path from one node to another node if and only if a skew-symmetric matrix pencil corresponding to the first node can be transformed by an arbitrarily small perturbation to a skew-symmetric matrix pencil corresponding to the second node. From the graph it is straightforward to identify more degenerate and more generic nearby canonical structures. A necessary (but not sufficient) condition for one orbit being in the closure of another is that the first orbit has larger codimension than the second one. Therefore we compute the codimensions of the congruence orbits (or bundles). It is done via the solutions of an associated homogeneous system of matrix equations. The complete stratification is done by proving the relation between equivalence and congruence for the skew-symmetric matrix pencils. This relation allows us to use the known result about the stratifications of general matrix pencils (under strict equivalence) in order to stratify skew-symmetric matrix pencils under congruence. Matlab functions to work with skew-symmetric matrix pencils and a number of other types of symmetries for matrices and matrix pencils are developed and included in the Matrix Canonical Structure (MCS) Toolbox.
303

A Drucker-Prager model for elastic contact with friction / A Drucker-Prager model for elastic contact with friction

wu, yunxian, wang, yiyun January 2011 (has links)
In mumerical contact simulations with friction, the simple Coloumb law is usually employed. Standard plasticity models are difficult to use since the balance enforced on the contact surface typically only involves balance of traction vectors, and does not use the full stress tensor on the interface. In this work we describe an approach that allows for the use of the stress tensor, thus opening up the possibility of using more advanced plasticity models. We exemplify this approach by implementing the Drucker-Prager pressure sensitive plasticity model.
304

Applying Autonomous Methods for Signal Analysis and Correction with Applications in the ship Industry

EL OUARDI, Abdelghafour January 2018 (has links)
The manufacturing and transportation industries generate a large amount of data sets which are often of inconsistent quality. The goal of this project is to find the mathematical principles of a system which learns automatically the essential statistical and analytical properties of datasets in order to detect and correct certain classes of faults in real time.
305

Evaluation of SPH for hydrodynamic modeling,using DualSPHysics

Eriksson, Jonas January 2018 (has links)
Computational methods are always being invented, improved and adjusted to newkinds of problems, this is a constant process happening all the time. The studyevaluates a method called Smoothed Particle Hydrodynamics (SPH) for modelingon fluid flows around ship hulls. This has been done mainly using a open sourcecode called DualSPHysics. The SPH method has been applied to complex problemsas well as simple problems for comparison to well known phenomena. It is aearly study of the method and aimed at discovering how to proceed when studyingthe method in the future. The results seem promising especially when computationsare made using Graphics Processing Units (GPU) for calculations. The codeDualSPHysics used in the study shows promise but might be in need of some morefunctions before being practically applicable for simulation of ship hulls.
306

Leveraging multicore processors for scientific computing

Tillenius, Martin January 2012 (has links)
This thesis deals with how to develop scientific computing software that runs efficiently on multicore processors. The goal is to find building blocks and programming models that increase the productivity and reduce the probability of programming errors when developing parallel software. In our search for new building blocks, we evaluate the use of hardware transactional memory for constructing atomic floating point operations. Using benchmark applications from scientific computing, we show in which situations this achieves better performance than other approaches. Driven by the needs of scientific computing applications, we develop a programming model and implement it as a reusable library. The library provides a run-time system for executing tasks on multicore architectures, with efficient and user-friendly management of dependencies. Our results from scientific computing benchmarks show excellent scaling up to at least 64 cores. We also investigate how the execution time depend on the task granularity, and build a model for the performance of the task library. / UPMARC / eSSENCE
307

Parallel algorithms and implementations for genetic analysis of quantitative traits

Jayawardena, Mahen January 2007 (has links)
Many important traits in plants, animals and humans are quantitative, and most such traits are generally believed to be regulated by multiple genetic loci. Standard computational tools for analysis of quantitative traits use linear regression models for relating the observed phenotypes to the genetic composition of individuals in a population. However, using these tools to simultaneously search for multiple genetic loci is very computationally demanding. The main reason for this is the complex nature of the optimization landscape for the multidimensional global optimization problems that must be solved. This thesis describes parallel algorithms and implementation techniques for such optimization problems. The new computational tools will eventually enable genetic analysis exploiting new classes of multidimensional statistical models, potentially resulting in interesting results in genetics. We first describe how the algorithm used for global optimization in the standard, serial software is parallelized and implemented on a grid system. Then, we also describe a parallelized version of the more elaborate global optimization algorithm DIRECT and show how this can be deployed on grid systems and other loosely-coupled architectures. The parallel DIRECT scheme is further developed to exploit both coarse-grained parallelism in grid or clusters as well as fine-grained, tightly-coupled parallelism in multi-core nodes. The results show that excellent speedup and performance can be archived on grid systems and clusters, even when using a tightly-coupled algorithms such as DIRECT. Finally, a pilot implementation of a grid portal providing a graphical front-end for our code is implemented. After some further development, this portal can be utilized by geneticists for performing multidimensional genetic analysis of quantitative traits on a regular basis.
308

Parallelizing the Method of Conjugate Gradients for Shared Memory Architectures

Löf, Henrik January 2004 (has links)
Solving Partial Differential Equations (PDEs) is an important problem in many fields of science and engineering. For most real-world problems modeled by PDEs, we can only approximate the solution using numerical methods. Many of these numerical methods result in very large systems of linear equations. A common way of solving these systems is to use an iterative solver such as the method of conjugate gradients. Furthermore, due to the size of these systems we often need parallel computers to be able to solve them in a reasonable amount of time. Shared memory architectures represent a class of parallel computer systems commonly used both in commercial applications and in scientific computing. To be able to provide cost-efficient computing solutions, shared memory architectures come in a large variety of configurations and sizes. From a programming point of view, we do not want to spend a lot of effort optimizing an application for a specific computer architecture. We want to find methods and principles of optimizing our programs that are generally applicable to a large class of architectures. In this thesis, we investigate how to implement the method of conjugate gradients efficiently on shared memory architectures. We seek algorithmic optimizations that result in efficient programs for a variety of architectures. To study this problem, we have implemented the method of conjugate gradients using OpenMP and we have measured the runtime performance of this solver on a variety of both uniform and non-uniform shared memory architectures. The input data used in the experiments come from a Finite-Element discretization of the Maxwell equations in three dimensions of a fighter-jet geometry. Our results show that, for all architectures studied, optimizations targeting the memory hierarchy exhibited the largest performance increase. Improving the load balance, by balancing the arithmetical work and minimizing the number of global barriers showed to be of lesser importance. Overall, bandwidth minimization of the iteration matrix showed to be the most efficient optimization. On non-uniform architectures, proper data distribution showed to be very important. In our experiments we used page migration to improve the data distribution during runtime. Our results indicate that page migration can be very efficient if we can keep the migration cost low. Furthermore, we believe that page migration can be introduced in a portable way into OpenMP in the form of a directive with a affinity-on-next-touch semantic.
309

Techniques for finite element methods on modern processors

Ljungkvist, Karl January 2015 (has links)
In this thesis, methods for efficient utilization of modern computer hardware for numerical simulation are considered. In particular, we study techniques for speeding up the execution of finite-element methods. One of the greatest challenges in finite-element computation is how to efficiently perform the the system matrix assembly efficiently in parallel, due to its complicated memory access pattern. The main difficulty lies in the fact that many entries of the matrix are being updated concurrently by several parallel threads. We consider transactional memory, an exotic hardware feature for concurrent update of shared variables, and conduct benchmarks on a prototype processor supporting it. Our experiments show that transactions can both simplify programming and provide good performance for concurrent updates of floating point data. Furthermore, we study a matrix-free approach to finite-element computation which avoids the matrix assembly. Motivated by its computational properties, we implement the matrix-free method for execution on graphics processors, using either atomic updates or a mesh coloring approach to handle the concurrent updates. A performance study shows that on the GPU, the matrix-free method is faster than a matrix-based implementation for many element types, and allows for solution of considerably larger problems. This suggests that the matrix-free method can speed up execution of large realistic simulations. / UPMARC / eSSENCE
310

Scientific computing on hybrid architectures

Holm, Marcus January 2013 (has links)
Modern computer architectures, with multicore CPUs and GPUs or other accelerators, make stronger demands than ever on writers of scientific code. As a rule of thumb, the fastest, most efficient program consists of labor-intensive code written by expert programmers for a certain application on a particular computer. This thesis deals with several algorithmic and technical approaches towards effectively satisfying the demand for high-performance parallel programming without incurring such a high cost in expert programmer time. Effective programming is accomplished by writing performance-portable code where performance-critical functionality is provided either by external software or at least a balance between maintainability/generality and efficiency. / UPMARC / eSSENCE

Page generated in 0.1007 seconds