Spelling suggestions: "subject:"high performance anda"" "subject:"high performance ando""
111 |
Computer simulations of polymers and gelsWood, Dean January 2013 (has links)
Computer simulations have become a vital tool in modern science. The ability to reliably move beyond the capabilities of experiment has allowed great insights into the nature of matter. To enable the study of a wide range of systems and properties a plethora of simulation techniques have been developed and refined, allowing many aspects of complex systems to be demystified. I have used a range of these to study a variety of systems, utilising the latest technology in high performance computing (HPC) and novel, nanoscale models. Monte Carlo (MC) simulation is a commonly used method to study the properties of system using statistical mechanics and I have made use of it in published work [1] to study the properties of ferrogels in homogeneous magnetic fields using a simple microscopic model. The main phenomena of interest concern the anisotropy and enhancement of the elastic moduli that result from applying uniform magnetic fields before and after the magnetic grains are locked in to the polymer-gel matrix by cross-linking reactions. The positional organization of the magnetic grains is influenced by the application of a magnetic field during gel formation, leading to a pronounced anisotropy in the mechanical response of the ferrogel to an applied magnetic field. In particular, the elastic moduli can be enhanced to different degrees depending on the mutual orientation of the fields during and after ferrogel formation. Previously, no microscopic models have been produced to shed light on this effect and the main purpose of the work presented here is to illuminate the microscopic behaviour. The model represents ferrogels by ensembles of dipolar spheres dispersed in elastic matrices. Experimental trends are shown to be reflected accurately in the simulations of the microscopic model while shedding light on the microscopic mechanism causing these effects. These mechanisms are shown to be related to the behaviour of the dipoles during the production of the gels and caused by the chaining of dipoles in magnetic fields. Finally, simple relationships between the elastic moduli and the magnetization are proposed. If supplemented by the magnetization curve, these relationships yield the dependencies of the elastic moduli on the applied magnetic field, which are often measured directly in experiments. While MC simulations are useful for statistical studies, it can be difficult to use them to gather information about the dynamics of a system. In this case, Molecular Dynamics (MD) is more widely used. MD generally utilises the classical equations of motion to simulate the evolution of a system. For large systems, which are often of interest, and multi-species polymers, the required computer power still poses a challenge and requires the use of HPC techniques. The most recent development in HPC is the use of Graphical Processing Units (GPU) for the fast solution of data parallel problems. In further published work [2], I have used a bespoke MD code utilising GPU acceleration in order to simulate large systems of block copolymers(BC) in solvent over long timescales. I have studied thin films of BC solutions drying on a flat, smooth surface which requires long timescales due to the ’slow’ nature of the process. BC’s display interesting self-organisation behaviour in bulk solution and near surfaces and have a wide range of potential applications from semi-conductors to self-constructing fabrics. Previous studies have shown some unusual behaviour of PI-PEO diblock co-polymers adsorbing to a freshly cleaved mica surface. These AFM studies showed polymers increasing in height over time and proposed the change of affinity of mica to water and the loss of water layers on the surface as a driver for this change. The MD simulation aimed to illuminate the process involved in this phenomena. The process of evaporation of water layers from a surface was successfully simulated and gave a good indication that the process of solvent evaporation from the surface and the ingress of solvent beneath the adsorbed polymer caused the increase in height seen in experiment.
|
112 |
Compilation techniques for high-performance embedded systems with multiple processorsFranke, Bjorn January 2004 (has links)
Despite the progress made in developing more advanced compilers for embedded systems, programming of embedded high-performance computing systems based on Digital Signal Processors (DSPs) is still a highly skilled manual task. This is true for single-processor systems, and even more for embedded systems based on multiple DSPs. Compilers often fail to optimise existing DSP codes written in C due to the employed programming style. Parallelisation is hampered by the complex multiple address space memory architecture, which can be found in most commercial multi-DSP configurations. This thesis develops an integrated optimisation and parallelisation strategy that can deal with low-level C codes and produces optimised parallel code for a homogeneous multi-DSP architecture with distributed physical memory and multiple logical address spaces. In a first step, low-level programming idioms are identified and recovered. This enables the application of high-level code and data transformations well-known in the field of scientific computing. Iterative feedback-driven search for “good” transformation sequences is being investigated. A novel approach to parallelisation based on a unified data and loop transformation framework is presented and evaluated. Performance optimisation is achieved through exploitation of data locality on the one hand, and utilisation of DSP-specific architectural features such as Direct Memory Access (DMA) transfers on the other hand. The proposed methodology is evaluated against two benchmark suites (DSPstone & UTDSP) and four different high-performance DSPs, one of which is part of a commercial four processor multi-DSP board also used for evaluation. Experiments confirm the effectiveness of the program recovery techniques as enablers of high-level transformations and automatic parallelisation. Source-to-source transformations of DSP codes yield an average speedup of 2.21 across four different DSP architectures. The parallelisation scheme is – in conjunction with a set of locality optimisations – able to produce linear and even super-linear speedups on a number of relevant DSP kernels and applications.
|
113 |
Development of methodology for high performance liquid chromatographicseparation of inorganic ions譚偉明, Tam, Wai-ming. January 1990 (has links)
published_or_final_version / Chemistry / Doctoral / Doctor of Philosophy
|
114 |
Performance studies of high-speed communication on commodity cluster譚達俊, Tam, Tat-chun, Anthony. January 2001 (has links)
published_or_final_version / Computer Science and Information Systems / Doctoral / Doctor of Philosophy
|
115 |
Parallel finite element analysisMargetts, Lee January 2002 (has links)
Finite element analysis is versatile and used widely in a range of engineering andscientific disciplines. As time passes, the problems that engineers and designers areexpected to solve are becoming more computationally demanding. Often theproblems involve the interplay of two or more processes which are physically andtherefore mathematically coupled. Although parallel computers have been availablefor about twenty years to satisfy this demand, finite element analysis is still largelyexecuted on serial machines. Parallelisation appears to be difficult, even for thespecialist. Parallel machines, programming languages, libraries and tools are used toparallelise old serial programs with mixed success. In some cases the serialalgorithm is not naturally suitable for parallel computing. Some argue that rewritingthe programs from scratch, using an entirely different solution strategy is a betterapproach. Taking this point of view, using MPI for portability, a mesh free elementby element method for simple data distribution and the appropriate iterative solvers,a general parallel strategy for finite element analysis is developed and assessed.
|
116 |
Exploring biodegradation of emerging pollutants using next generation sequencing and UPLC-MS-MS techniquesYu, Ke, 余珂 January 2014 (has links)
This study was conducted to set up a systematic approach utilizing advantages of both wet lab and bioinformatic methodologies to study biodegradation abilities and microbial bacterial-functional relationship within bioremediation communities.
Firstly, 11pharmaceuticals and personal care products (PPCPs)were selected as target chemicals for establishing an effective determination process in analyzing trace-level concentrations in the environment, and understanding the removal routes during pollutants removal process in wastewater treatment process using activated sludge. Ultra performance liquid chromatography-tandem mass spectrometry was utilized to develop a rapid, sensitive and reliable method without solid phase extraction pre-concentration for trace analysis of 11 PPCPs in influent and effluent from municipal wastewater treatment plants. Shorten the detection time and significant reduction of detection cost were achieved due to the omitting usage of solid phase extraction (SPE)process and avoiding the consumption of hydrophiliclipophilic balancced (HLB)cartridge.
Research on removal routes of ten selected PPCPs in activated sludge found activated sludge hardly removed carbamazepine. Biodegradation was the sole route to remove acyclovir, metronidazole, benzylparaben, ethylparaben, methylparaben and propylparaben. Both adsorption and biodegradation were involved in the removal of ranitidine and benzophenone-3, while fluoxetine could be totally removed by adsorption in activated sludge.
Secondly, as the target microbial community, activated sludge community was used to set up the global bioinformatic analysis process. Both metagenomic and metatranscriptomic approaches were processed to characterize microbial structure and gene expression of activated sludge community. The taxonomic profile showed thatactivated sludge was dominated by Proteobacteria, Actinobacteria, Bacteroidetes, Firmicutes and Verrucomicrobiaphyla. Gene expression annotation of nitrogen removal revealed that denitrification-related genes sequences dominated in both DNA and cDNA datasets while nitrifying genes were also expressed in relative high levels. Specially, ammonia monooxygenase and hydroxylamine oxidase demonstrated the high cDNA/DNA ratios, indicating strong nitrification activity. Ammonia-oxidizing bacteria present mainly belonged to Nitrosomonas and Nitrosospira species.
A fast method to construct local sub-databases has been established for the quick similarity search and annotation of huge metagenomic datasets. The conducted tests showed sub-database annotation pipeline achieved a speedup of ~150-385 times, and got exactly the same annotation results with those of the direct NCBI-nr database BLAST-MEGAN method. This approach provides a new time-efficient and convenient annotation similarity search strategy for laboratories without access to high performance computing facilities.
Thirdly, bisphenol A(BPA), which has a partially known biodegradation pathway and relevant bioremediating genes, was chosen as a model to establish a pipeline for systematical understanding the pathways and gene/bacteria relationships in an enriched microbial community. 11 new metabolites were detected during BPA degradation. Thereby, a novel pathway of degrading BPA metabolite was proposed. Sphingomonas strains were dominant taxa in initial degradation of BPA, while the other taxa were competing BPA metabolites during degradation. Metagenomic binning results showed a cytochrome P450 monooxygenase system, which was previously reported BPA mediator, was sharing by two Sphingomonas strains, showing the undergoing mechanism of competition of the two strains. The observations suggested bacterial specialization may occur in that community that each taxon was selected to degrade certain metabolite in a community economical way. / published_or_final_version / Civil Engineering / Doctoral / Doctor of Philosophy
|
117 |
Energy efficient branch predictionHicks, Michael Andrew January 2010 (has links)
Energy efficiency is of the utmost importance in modern high-performance embedded processor design. As the number of transistors on a chip continues to increase each year, and processor logic becomes ever more complex, the dynamic switching power cost of running such processors increases. The continual progression in fabrication processes brings a reduction in the feature size of the transistor structures on chips with each new technology generation. This reduction in size increases the significance of leakage power (a constant drain that is proportional to the number of transistors). Particularly in embedded devices, the proportion of an electronic product’s power budget accounted for by the CPU is significant (often as much as 50%). Dynamic branch prediction is a hardware mechanism used to forecast the direction, and target address, of branch instructions. This is essential to high performance pipelined and superscalar processors, where the direction and target of branches is not computed until several stages into the pipeline. Accurate branch prediction also acts to increase energy efficiency by reducing the amount of time spent executing mis-speculated instructions. ‘Stalling’ is no longer a sensible option when the significance of static power dissipation is considered. Dynamic branch prediction logic typically accounts for over 10% of a processor’s global power dissipation, making it an obvious target for energy optimisation. Previous approaches at increasing the energy efficiency of dynamic branch prediction logic has focused on either fully dynamic or fully static techniques. Dynamic techniques include the introduction of a new cache-like structure that can decide whether branch prediction logic should be accessed for a given branch, and static techniques tend to focus on scheduling around branch instructions so that a prediction is not needed (or the branch is removed completely). This dissertation explores a method of combining static techniques and profiling information with simple hardware support in order to reduce the number of accesses made to a branch predictor. The local delay region is used on unconditional absolute branches to avoid prediction, and, for most other branches, Adaptive Branch Bias Measurement (through profiling) is used to assign a static prediction that is as accurate as a dynamic prediction for that branch. This information is represented as two hint-bits in branch instructions, and then interpreted by simple hardware logic that bypasses both the lookup and update phases for appropriate branches. The global processor power saving that can be achieved by this Combined Algorithm is around 6% on the experimental architectures shown. These architectures are based upon real contemporary embedded architecture specifications. The introduction of the Combined Algorithm also significantly reduces the execution time of programs on Multiple Instruction Issue processors. This is attributed to the increase achieved in global prediction accuracy.
|
118 |
Establishing Linux Clusters for high-performance computing (HPC) at NPSDaillidis, Christos 09 1900 (has links)
Approved for public release; distribution is unlimited / S tasks. Discrete Event Simulation (DES) often involves repeated, independent runs of the same models with different input parameters. A system which is able to run many replications quickly is more useful than one in which a single monolithic application runs quickly. A loosely coupled parallel system is indicated. Inexpensive commodity hardware, high speed local area networking, and open source software have created the potential to create just such loosely coupled parallel systems. These systems are constructed from Linux-based computers and are called Beowulf clusters. This thesis presents an analysis of clusters in high-performance computing and establishes a testbed implementation at the MOVES Institute. It describes the steps necessary to create a cluster, factors to consider in selecting hardware and software, and describes the process of creating applications that can run on the cluster. Monitoring the running cluster and system administration are also addressed. / Major, Hellenic Army
|
119 |
A comparative study of ethanolic versus triturated dilutions in terms of the amount of caffeine extracted from Coffea tosta by means of high pressure liquid chromatographyHarris, Bronwyn Claire January 2002 (has links)
A mini-dissertation in partial compliance with the requirements for a Master's Degree in Technology: Homoeopathy, Durban Institute of Technology, 2002. / The purpose of this study was to compare the amount of caffeine extracted from triturated samples and ethanolic samples of Coffea tosta using high pressure liquid chromatography (HPLC) as a method of analysis. The study wanted to expand on homoeopathic pharmaceutical knowledge, specifically looking at the two methods of remedy preparation of plant materials. From the same batch of ground roasted coffee beans, using the decimal scale of dilution, the mother tincture (bill) and the first triturated (bill) samples were prepared. The subsequent 2xH and 3xH triturated and ethanolic potencies were then made in accordance with homoeopathic methodology. Each group contained three different dilution levels (bill, 2xH and 3xH), 18 samples per group giving a total of36 samples that were analysed using HPLC. Three samples were analysed from the three dilution levels in each Group, in total there were 18 samples from the triturated group and 18 from the ethanolic group. . The samples were analysed quantitatively using the highly accurate and advanced method of high pressure liquid chromatography. This method gives accurate readings of the caffeine concentrations of a sample compared to a caffeine standard. This allowed for quantification of the caffeine concentration of each sample. The percentage caffeine was calculated from each sample. The aim of the study was to evaluate the difference in each method of preparation by measuring the amount of caffeine extracted from the samples. The results obtained from the inter-Group Mann-Whitney and ANOVA tests showed that there was a significant difference between the ethanolic dilutions and triturated dilutions with regards to the 1xH and 2xH dilutions. In the 1xH dilution the ethanolic method retained / M
|
120 |
Drilling of high-performance materials: experimental, numerical, and theoretical investigationsCong, Weilong January 1900 (has links)
Doctor of Philosophy / Department of Industrial & Manufacturing Systems Engineering / Zhijian Pei / High-performance materials, such as silicon, aerospace stainless steels, titanium alloys, and carbon fiber reinforced plastic (CFRP) composites, have a variety of engineering applications. However, they usually have poor machinability and are classified as hard-to-machine materials. Drilling is one of the important machining processes for these materials. Industries are always under tremendous pressure to meet the ever-increasing demand for lower cost and better quality of the products made from these high-performance materials.
Rotary ultrasonic machining (RUM) is a non-traditional machining process that combines the material removal mechanisms of diamond grinding and ultrasonic machining. It is a relatively low-cost, environment-benign process that easily fits in the infrastructure of the traditional machining environment. Other advantages of this process include high hole accuracy, superior surface finish, high material removal rate, low tool pressure, and low tool wear rate.
The goal of this research is to provide new knowledge of machining these high performance materials with RUM for further improvement in the machined hole quality and decrease in the machining cost. A thorough research in this dissertation has been conducted by experimental, numerical, and theoretical investigations on output variables, including cutting force, torque, surface roughness, tool wear, cutting temperature, material removal rate, edge chipping (for silicon), power consumption (for CFRP), delamination (for CFRP), and feasible regions (for dry machining of CFRP). In this dissertation, an introduction of workpiece materials and RUM are discussed first. After that, two literature reviews on silicon drilling and dry drilling are presented. Then, design of experiment and finite element analysis on edge chipping in RUM of silicon, experimental investigations and finite element analysis on RUM of aerospace stainless steels, an ultrasonic vibration amplitude measurement method and a cutting temperature measurement method for RUM using titanium alloys as workpiece, experimental and theoretical investigations on RUM of CFRP composites, and experimental studies on CFRP/Ti stacks are presented, respectively. Finally, conclusions and contributions on RUM drilling are discussed.
|
Page generated in 0.0975 seconds