• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 337
  • 189
  • 134
  • 56
  • 45
  • 44
  • 4
  • 4
  • 2
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 922
  • 922
  • 922
  • 404
  • 394
  • 351
  • 351
  • 329
  • 325
  • 320
  • 319
  • 316
  • 314
  • 313
  • 313
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
361

A Parallel Adaptive Mesh Refinement Library for Cartesian Meshes

January 2019 (has links)
abstract: This dissertation introduces FARCOM (Fortran Adaptive Refiner for Cartesian Orthogonal Meshes), a new general library for adaptive mesh refinement (AMR) based on an unstructured hexahedral mesh framework. As a result of the underlying unstructured formulation, the refinement and coarsening operators of the library operate on a single-cell basis and perform in-situ replacement of old mesh elements. This approach allows for h-refinement without the memory and computational expense of calculating masked coarse grid cells, as is done in traditional patch-based AMR approaches, and enables unstructured flow solvers to have access to the automated domain generation capabilities usually only found in tree AMR formulations. The library is written to let the user determine where to refine and coarsen through custom refinement selector functions for static mesh generation and dynamic mesh refinement, and can handle smooth fields (such as level sets) or localized markers (e.g. density gradients). The library was parallelized with the use of the Zoltan graph-partitioning library, which provides interfaces to both a graph partitioner (PT-Scotch) and a partitioner based on Hilbert space-filling curves. The partitioned adjacency graph, mesh data, and solution variable data is then packed and distributed across all MPI ranks in the simulation, which then regenerate the mesh, generate domain decomposition ghost cells, and create communication caches. Scalability runs were performed using a Leveque wave propagation scheme for solving the Euler equations. The results of simulations on up to 1536 cores indicate that the parallel performance is highly dependent on the graph partitioner being used, and differences between the partitioners were analyzed. FARCOM is found to have better performance if each MPI rank has more than 60,000 cells. / Dissertation/Thesis / Doctoral Dissertation Aerospace Engineering 2019
362

PPerfGrid: A Grid Services-Based Tool for the Exchange of Heterogeneous Parallel Performance Data

Hoffman, John Jared 01 January 2004 (has links)
This thesis details the approach taken in developing PPerfGrid. Section 2 discusses other research related to this project. Section 3 provides general background on the technologies utilized in PPerfGrid, focusing on the components that make up the Grid services architecture. Section 4 provides a description of the architecture of PPerfGrid. Section 5 details the implementation of PPerfGrid. Section 6 presents tests designed to measure the overhead and scalability of the PPerfGrid application. Section 7 suggests future work, and Section 8 concludes the thesis.
363

Fast-NetMF: Graph Embedding Generation on Single GPU and Multi-core CPUs with NetMF

Shanmugam Sakthivadivel, Saravanakumar 24 October 2019 (has links)
No description available.
364

MARS: Multi-Scalable Actor-Critic Reinforcement Learning Scheduler

Baheri, Betis 24 July 2020 (has links)
No description available.
365

A novel mesh generator for the numerical simulation of multi-scale physics in neurons

Grein, Stephan, 0000-0001-9524-6633 January 2020 (has links)
Computational Neuroscience deals with spatio-temporal scales which vary considerably.For example interactions at synaptic contact regions occur on the scale of nanometers and nanoseconds to milliseconds (micro-scale) whereas networks of neurons can measure up to millimeters and signals are processed on the scale of seconds (macro-scale). Whole-cell calcium dynamics models (meso-scale) mediate between the multiple spatio-temporal scales. Of crucial importance is the calcium propagation mediated by the highly complex endoplasmic reticulum network. Most models do not account for the intricate intracellular architecture of neurons and consequently cannot resolve the interplay between structure and calcium-mediated function. To incorporate the detailed cellular architecture in intracellular Calcium models, a novel mesh generation methodology has been developed to allow for the efficient generation of computational meshes of neurons with a three-dimensionally resolved endoplasmic reticulum. Mesh generation routines are compiled into a versatile and fully automated reconstruct-and-simulation toolbox for multi-scale physics to be utilized on high-performance or regular computing infrastructures. First-principle numerical simulations on the neuronal reconstructions reveal that intracellular Calcium dynamics are effected by morphological features of the neurons, for instance a change of endoplasmic reticulum diameter leads to a significant spatio-temporal variability of the calcium signal at the soma. / Math & Science Education
366

High Performance Data Mining Techniques For Intrusion Detection

Siddiqui, Muazzam Ahmed 01 January 2004 (has links)
The rapid growth of computers transformed the way in which information and data was stored. With this new paradigm of data access, comes the threat of this information being exposed to unauthorized and unintended users. Many systems have been developed which scrutinize the data for a deviation from the normal behavior of a user or system, or search for a known signature within the data. These systems are termed as Intrusion Detection Systems (IDS). These systems employ different techniques varying from statistical methods to machine learning algorithms. Intrusion detection systems use audit data generated by operating systems, application softwares or network devices. These sources produce huge amount of datasets with tens of millions of records in them. To analyze this data, data mining is used which is a process to dig useful patterns from a large bulk of information. A major obstacle in the process is that the traditional data mining and learning algorithms are overwhelmed by the bulk volume and complexity of available data. This makes these algorithms impractical for time critical tasks like intrusion detection because of the large execution time. Our approach towards this issue makes use of high performance data mining techniques to expedite the process by exploiting the parallelism in the existing data mining algorithms and the underlying hardware. We will show that how high performance and parallel computing can be used to scale the data mining algorithms to handle large datasets, allowing the data mining component to search a much larger set of patterns and models than traditional computational platforms and algorithms would allow. We develop parallel data mining algorithms by parallelizing existing machine learning techniques using cluster computing. These algorithms include parallel backpropagation and parallel fuzzy ARTMAP neural networks. We evaluate the performances of the developed models in terms of speedup over traditional algorithms, prediction rate and false alarm rate. Our results showed that the traditional backpropagation and fuzzy ARTMAP algorithms can benefit from high performance computing techniques which make them well suited for time critical tasks like intrusion detection.
367

Efficient and Scalable Simulations of Active Hydrodynamics in Three Dimensions

Singh, Abhinav 14 February 2024 (has links)
Active matter represents a unique class of non-equilibrium systems, including examples ranging from cellular structures to large-scale biological tissues. These systems exhibit intriguing spatiotemporal dynamics, driven by the constituent particles’ continuous energy expenditure. Such active-matter systems, featuring complex hydrodynamics, are described by sophisticated mathematical models, typically using partial differential equations (PDEs). PDEs modeling hydrodynamics, such as the Navier-Stokes equations, are analytically intractable, and notoriously challenging to study computationally. The challenges include the need for consistent numerical methods along with their efficient and scalable high-performance computer implementation to solve the PDEs numerically. However, when considering new theoretical PDE models, such as active hydrodynamics, conventional approaches often fall short due to the specialization made in the numerical methods to study certain specific models. The inherent complexity and nonlinearity of active-matter PDEs add to the challenge. Hence, the computational study of such active-matter PDE models requires rapidly evolving high-performance computer software that can easily implement new numerical methods to solve these equations in biologically realistic three-dimensional domains. This presents a rich, yet underexplored territory demanding scalable computational frameworks that apply to a large class of PDEs. In this thesis, we introduce a computational framework that effectively allows for using multiple numerical methods through a context-aware template expression system akin to an embedded domain-specific language. This framework primarily aims at solving lengthy PDEs associated with active hydrodynamics in complex domains, while experimenting with new numerical methods. Existing PDE-solving codes often lack this flexibility, as they are closely tied to a PDE and domain geometry that rely on a specific numerical method. We overcome these limitations by using an object-oriented implementation design, and show experiments with adaptive and numerically consistent particle-based approach called Discretization-Corrected Particle Strength Exchange (DC-PSE). DC-PSE allows for the higher-order discretization of differential operators on arbitrary particle distributions leading to the possibility of solving active hydrodynamic PDEs in complex domains. However, the curse of dimensionality makes it difficult to numerically solve three-dimensional equations on single-core architectures and warrants the use of parallel and distributed computers. We design a novel template-expression system and implement it in the scalable scientific computing library OpenFPM. Our methodology offers an expression-based embedded language, enabling PDE codes to be written in a form that closely mirrors mathematical notation. Leveraging OpenFPM, this approach also ensures parallel scalability. To further enhance our framework's versatility, we employ a \textit{separation-of-concerns} abstraction, segregating the model equations from numerics, and domain geometry. This allows for the rapid rewriting of codes for agile numerical experiments across different model equations in various geometries. Supplementing this framework, we develop a distributed algebra system compatible with OpenFPM and Boost Odeint. This algebra system opens avenues for a multitude of explicit adaptive time-integration schemes, which can be selected by modifying a single line of code while maintaining parallel scalability. Motivated by symmetry-preserving theories of active hydrodynamics, and as a first benchmark of our template-expression system, we present a high-order numerically convergent scheme to study active polar fluids in arbitrary three-dimensional domains. We derive analytical solutions in simple Cartesian geometries and use them to show the numerical convergence of our algorithm. Further, we showcase the scalability of the computer code written using our expression system on distributed computing systems. To cater to the need for solving PDEs on curved surfaces, we present a novel meshfree numerical scheme, the Surface DC-PSE method. Upon implementation in our scalable framework, we benchmark Surface DC-PSE for both explicit and implicit Laplace-Beltrami operators and show applications to computing mean and Gauss curvature. Finally, we apply our computational framework to exploring the three-dimensional active hydrodynamics of biological flowing matter, a prominent model system to study the active dynamics of cytoskeletal networks, celluar migration, and tissue mechanics. Our software framework effectively tackles the challenges associated to numerically solving such non-equilibrium spatiotemporal PDEs. We perform linear perturbation analysis of the three-dimensional Ericksen-Leslie model and find an analytical expression for the critical active potential or, equivalently, a critical length of the system above which a spontaneous flow transition occurs. This spontaneous flow transition is a first realization of a three-dimensional active Fr\'eedericksz transition. With our expression system, we successfully simulate 3D active fluids, finding phases of spontaneous flow transitions, traveling waves, and spatiotemporal chaos with increasing active stress. We numerically find a topological phase transition similar to the Berezinskii–Kosterlitz–Thouless transition (BKT transition) of the two-dimensional XY model that occurs in active polar fluids after the spontaneous flow transition. We then proceed to non-Cartesian geometries and show the application of our software framework to solve the active polar fluid equations in spherical domains. We find spontaneous flows in agreement with recent experimental observations. We further showcase the framework to solve the equations in 3D annular domains and a `peanut' geometry that resembles a dividing cell. Our simulations further recapitulate the actin flows observed in \textit egg extracts within spherical shell geometries, showcasing our framework's versatility in handling complex geometrical modifications of model equations. Looking ahead, we hope our framework will serve as a foundation for further advancements in computational morphogenesis, fostering collaboration and using the present techniques in biophysical modeling.
368

Design, development and evaluation of the ruggedized edge computing node (RECON)

Patel, Sahil Girin 09 December 2022 (has links)
The increased quality and quantity of sensors provide an ever-increasing capability to collect large quantities of high-quality data in the field. Research devoted to translating that data is progressing rapidly; however, translating field data into usable information can require high performance computing capabilities. While high performance computing (HPC) resources are available in centralized facilities, bandwidth, latency, security and other limitations inherent to edge location in field sensor applications may prevent HPC resources from being used in a timely fashion necessary for potential United States Army Corps of Engineers (USACE) field applications. To address these limitations, the design requirements for RECON are established and derived from a review of edge computing, in order to develop and evaluate a novel high-power, field-deployable HPC platform capable of operating in austere environments at the edge.
369

A Shared-Memory Coupled Architecture to Leverage Big Data Frameworks in Prototyping and In-Situ Analytics for Data Intensive Scientific Workflows

Lemon, Alexander Michael 01 July 2019 (has links)
There is a pressing need for creative new data analysis methods whichcan sift through scientific simulation data and produce meaningfulresults. The types of analyses and the amount of data handled by currentmethods are still quite restricted, and new methods could providescientists with a large productivity boost. New methods could be simpleto develop in big data processing systems such as Apache Spark, which isdesigned to process many input files in parallel while treating themlogically as one large dataset. This distributed model, combined withthe large number of analysis libraries created for the platform, makesSpark ideal for processing simulation output.Unfortunately, the filesystem becomes a major bottleneck in any workflowthat uses Spark in such a fashion. Faster transports are notintrinsically supported by Spark, and its interface almost denies thepossibility of maintainable third-party extensions. By leveraging thesemantics of Scala and Spark's recent scheduler upgrades, we forceco-location of Spark executors with simulation processes and enable fastlocal inter-process communication through shared memory. This provides apath for bulk data transfer into the Java Virtual Machine, removing thecurrent Spark ingestion bottleneck.Besides showing that our system makes this transfer feasible, we alsodemonstrate a proof-of-concept system integrating traditional HPC codeswith bleeding-edge analytics libraries. This provides scientists withguidance on how to apply our libraries to gain a new and powerful toolfor developing new analysis techniques in large scientific simulationpipelines.
370

In Situ Visualization of Performance Data in Parallel CFD Applications

Falcao do Couto Alves, Rigel 19 January 2023 (has links)
This thesis summarizes the work of the author on visualization of performance data in parallel Computational Fluid Dynamics (CFD) simulations. Current performance analysis tools are unable to show their data on top of complex simulation geometries (e.g. an aircraft engine). But in CFD simulations, performance is expected to be affected by the computations being carried out, which in turn are tightly related to the underlying computational grid. Therefore it is imperative that performance data is visualized on top of the same computational geometry which they originate from. However, performance tools have no native knowledge of the underlying mesh of the simulation. This scientific gap can be filled by merging the branches of HPC performance analysis and in situ visualization of CFD simulations data, which shall be done by integrating existing, well established state-of-the-art tools from each field. In this threshold, an extension for the open-source performance tool Score-P was designed and developed, which intercepts an arbitrary number of manually selected code regions (mostly functions) and send their respective measurements – amount of executions and cumulative time spent – to the visualization software ParaView – through its in situ library, Catalyst –, as if they were any other flow-related variable. Subsequently the tool was extended with the capacity to also show communication data (messages sent between MPI ranks) on top of the CFD mesh. Testing and evaluation are done with two industry-grade codes: Rolls-Royce’s CFD code, Hydra, and Onera, DLR and Airbus’ CFD code, CODA. On the other hand, it has been also noticed that the current performance tools have limited capacity of displaying their data on top of three-dimensional, framed (i.e. time-stepped) representations of the cluster’s topology. Parallel to that, in order for the approach not to be limited to codes which already have the in situ adapter, it was extended to take the performance data and display it – also in codes without in situ – on a three-dimensional, framed representation of the hardware resources being used by the simulation. Testing is done with the Multi-Grid and Block Tri-diagonal NAS Parallel Benchmarks (NPB), as well as with Hydra and CODA again. The benchmarks are used to explain how the new visualizations work, while real performance analyses are done with the industry-grade CFD codes. The proposed solution is able to provide concrete performance insights, which would not have been reached with the current performance tools and which motivated beneficial changes in the respective source code in real life. Finally, its overhead is discussed and proven to be suitable for usage with CFD codes. The dissertation provides a valuable addition to the state of the art of highly parallel CFD performance analysis and serves as basis for further suggested research directions.

Page generated in 0.536 seconds