• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 14
  • 5
  • 5
  • 2
  • 1
  • Tagged with
  • 31
  • 31
  • 9
  • 7
  • 7
  • 4
  • 4
  • 3
  • 3
  • 3
  • 3
  • 3
  • 3
  • 3
  • 2
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

An experimental and numerical study of an automotive cooling module

Rynell, Anders January 2017 (has links)
Heavy vehicles are major emitters of noise. Especially at idle or low vehicle speeds a large portion of the noise emanates from the fan that forces the flow through the cooling module. The aim of this work is to investigate and reveal aerodynamic and acoustic installation effects linked to the cooling package. This introduces a multidisciplinary approach involving examination of the flow field, sound generation and sound propagation. The work includes two main parts: an experimental and a numerical part. The cooling module used throughout this work, named reduced cooling module, primarily includes a radiator, a shroud, a fan and a hydraulic engine to simplify the aeroacoustics analysis. The experimental part comprises measurements of the sound emanated from the cooling package. A new approach to the spectral decomposition method is developed yielding the fan sound power or spectrum to be formulated as a product of a source part and a system part scaling with the Strouhal number and the Helmholtz number. Also, a separate determination of the transmission loss of the radiator is performed. The impact of the radiator on the transmitted noise was found to be negligible. The numerical part incorporates comparisons from two aeroacoustics studies; a configuration where the fan is forced to operate at a fixed operation point and measured flow and turbulence statistics are available and the reduced cooling module. A hybrid turbulence modeling technique, IDDES, is adopted for the flow simulations. The sound propagation is calculated by the Ffowcs-Williams and Hawkings acoustic analogy when assuming a free-field sound propagation and by a finite element solver in the frequency domain to capture the installation effects. The simulated SPL conforms to the measured SPL and the blade response to the turbulent inflow and to the tip resolution, respectively, produce noise which spectral shape distribution is modified in accordance with earlier experimental findings published. Furthermore, the influence of an upstream radiator in close contact with the fan on the flow and sound fields is investigated. Here, the simulated aeroacoustic characteristics were found to change similarly to the acoustic measurements with and without radiator.
2

Spectral Decomposition Using S-transform for Hydrocarbon Detection and Filtering

Zhang, Zhao 2011 August 1900 (has links)
Spectral decomposition is a modern tool that utilizes seismic data to generate additional useful information in seismic exploration for hydrocarbon detection, lithology identification, stratigraphic interpretation, filtering and others. Different spectral decomposition methods with applications to seismic data were reported and investigated in past years. Many methods usually do not consider the non-stationary features of seismic data and, therefore, are not likely to give satisfactory results. S-transform developed in recent years is able to provide time-dependent frequency analysis while maintaining a direct relationship with the Fourier spectrum, a unique property that other methods of spectral decomposition may not have. In this thesis, I investigated the feasibility and efficiency of using S-transform for hydrocarbon detection and time-varying surface wave filtering. S-transform was first applied to two seismic data sets from a clastic reservoir in the North Sea and a deep carbonate reservoir in the Sichuan Basin, China. Results from both cases demonstrated that S-transform decomposition technique can detect hydrocarbon zones effectively and helps to build the relationships between lithology changes and high frequency variation and between hydrocarbon occurrence and low-frequency anomaly. However, its time resolution needs to be improved. In the second part of my thesis, I used S-transform to develop a novel Time-frequency-wave-number-domain (T-F-K) filtering method to separate surface wave from reflected waves in seismic records. The S-T-F-K filtering proposed here can be used to analyze surface waves on separate f-k panels at different times. The method was tested using hydrophone records of four-component seismic data acquired in the shallow-water Persian Gulf where the average water depth is about 10m and Scholte waves and other surfaces wave persistently strong. Results showed that this new S-T-F-K method is able to separate and sttenuate surface waves and to improve greatly the quality of seismic reflection signals that are otherwise completely concealed by the aliased surface waves.
3

Contributions to watermarking of 3D meshes/Contributions au tatouage des maillages surfaciques 3D

Cayre, François 09 December 2003 (has links)
We present two watermarking schemes for 3D meshes : - watermarking with geometrical invariant for fragile watermarking towards authentication and integrity purposes - watermarking in the geometrical spectral domain towards robust watermarking / Nous présentons deux schémas de tatouage pour maillages surfaciques 3D : - tatouage fragile par invariants géométriques pour l'authentification et l'intégrité - tatouage robuste dans l'espace de la décomposition spectrale
4

Frequency dependent seismic reflection analysis: a path to new direct hydrocarbon indicators for deep water reservoirs

Yoo, Seung Chul 02 June 2009 (has links)
To better study frequency related effects such as attenuation and tuning, we developed a frequency dependent seismic reflection analysis. Comprehensive tests on full waveform synthetics and observations from the Teal South ocean bottom seismic (OBS) data set confirmed that normal moveout (NMO) stretch could distort both frequency and amplitude information severely in shallow events and far offset traces. In synthetic tests, our algorithm recovered amplitude and frequency information ac-curately. This simple but robust target oriented NMO stretch correction scheme can be used on top of an existing seismic processing flow for further analyses. By combining the NMO stretch correction, spectral decomposition, and crossplots of am-plitude versus offset (AVO) attributes, we tested the frequency dependent workflow over Teal south and Ursa field data sets for improved reservoir characterization. As expected from NMO stretch characteristics, low frequencies have been less affected while mid and high frequency ranges were affected considerably. In seismic attribute analysis, the AVO crossplots from spectrally decomposed prestack data confirmed the improved accuracy and effectiveness of our workflow in mid and high frequency regions. To overcome poor spectral decomposition results due to low signal to noise ratio (S/N) in the Teal South application, we also implemented a substack scheme that stacks adjacent traces to increase S/N ratio while reducing the amount of data to process and increasing the accuracy of the spectral decomposition step. Synthetic tests verified the effectiveness of this additional step. An application to the Ursa, Gulf of Mexico, deep water data set showed significant improvement in high frequency data while correcting biased low frequency information.
5

Frequency dependent seismic reflection analysis: a path to new direct hydrocarbon indicators for deep water reservoirs

Yoo, Seung Chul 02 June 2009 (has links)
To better study frequency related effects such as attenuation and tuning, we developed a frequency dependent seismic reflection analysis. Comprehensive tests on full waveform synthetics and observations from the Teal South ocean bottom seismic (OBS) data set confirmed that normal moveout (NMO) stretch could distort both frequency and amplitude information severely in shallow events and far offset traces. In synthetic tests, our algorithm recovered amplitude and frequency information ac-curately. This simple but robust target oriented NMO stretch correction scheme can be used on top of an existing seismic processing flow for further analyses. By combining the NMO stretch correction, spectral decomposition, and crossplots of am-plitude versus offset (AVO) attributes, we tested the frequency dependent workflow over Teal south and Ursa field data sets for improved reservoir characterization. As expected from NMO stretch characteristics, low frequencies have been less affected while mid and high frequency ranges were affected considerably. In seismic attribute analysis, the AVO crossplots from spectrally decomposed prestack data confirmed the improved accuracy and effectiveness of our workflow in mid and high frequency regions. To overcome poor spectral decomposition results due to low signal to noise ratio (S/N) in the Teal South application, we also implemented a substack scheme that stacks adjacent traces to increase S/N ratio while reducing the amount of data to process and increasing the accuracy of the spectral decomposition step. Synthetic tests verified the effectiveness of this additional step. An application to the Ursa, Gulf of Mexico, deep water data set showed significant improvement in high frequency data while correcting biased low frequency information.
6

Investigation of time-lapse 4D seismic tuning and spectral responses to CO₂-EOR for enhanced characterization and monitoring of a thin carbonate reservoir

Krehel, Austin January 1900 (has links)
Master of Science / Department of Geology / Abdelmoneam Raef / Advancements, applications, and success of time-lapse (4D) seismic monitoring of carbonate reservoirs is limited by these systems’ inherent heterogeneity and low compressibility relative to siliciclastic systems. To contribute to the advancement of 4D seismic monitoring in carbonates, an investigation of amplitude envelope across frequency sub-bands was conducted on a high-resolution 4D seismic data set acquired in fine temporal intervals between a baseline and eight monitor surveys to track CO₂-EOR from 2003-2005 in the Hall-Gurney Field, Kansas. The shallow (approximately 900 m) Plattsburg ‘C Zone’ target reservoir is an oomoldic limestone within the Lansing-Kansas City (LKC) supergroup – deposited as a sequence of high-frequency, stacked cyclothems. The LKC reservoir fluctuates around thin-bed thickness within the well pattern region and is susceptible to amplitude tuning effects, in which CO₂ replacement of initial reservoir fluid generates a complex tuning phenomena with reduction and brightening of amplitude at reservoir thickness above and below thin-bed thickness, respectively. A thorough analysis of horizon snapping criteria and parameters was conducted to understand the sensitivity of these autonomous operations and produce a robust horizon tracking workflow to extend the Baseline Survey horizon data to subsequent Monitor Surveys. This 4D seismic horizon tracking workflow expedited the horizon tracking process across monitor surveys, while following a quantitative, repeatable approach in tracking the LKC and maintaining geologic integrity despite low signal-to-noise ratio (SNR) data and misties between surveys. Analysis of amplitude envelope data across frequency sub-bands (30-80 Hz) following spectral decomposition identified geometric features of multiple LKC shoal bodies at the reservoir interval. In corroboration with prior geologic interpretation, shoal boundaries, zones of overlap between stacked shoals, thickness variation, and lateral changes in lithofacies were delineated in the Baseline Survey, which enhanced detail of these features’ extent beyond capacity offered from well log data. Lineaments dominated by low-frequency anomalies within regions of adjacent shoals’ boundaries suggest thicker zones of potential shoal overlap. Analysis of frequency band-to-band analysis reveals relative thickness variation. Spectral decomposition of the amplitude envelope was analyzed between the Baseline and Monitor Surveys to identify spectral and tuning changes to monitor CO₂ migration. Ambiguity of CO₂ effects on tuning phenomena was observed in zones of known CO₂ fluid replacement. A series of lineaments highlighted by amplitude brightening from the Baseline to Monitor Surveys is observed, which compete with a more spatially extensive effect of subtle amplitude dimming. These lineaments are suggestive of features below tuning thickness, such as stratigraphic structures of shoals, fractures, and/or thin shoal edges, which are highlighted by an increased apparent thickness and onset of tuning from CO₂. Detailed analysis of these 4D seismic data across frequency sub-bands provide enhanced interpretation of shoal geometry, position, and overlap; identification of lateral changes in lithofacies suggestive of barriers and conduits; insight into relative thickness variation; and the ability of CO₂ tuning ambiguity to highlight zones below tuning thickness and improve reservoir characterization. These results suggest improved efficiency of CO₂ -EOR reservoir surveillance in carbonates, with implications to ensure optimal field planning and flood performance for analogous targets.
7

Perception and re-synchronization issues for the watermarking of 3D shapes

Rondao Alface, Patrice 26 October 2006 (has links)
Digital watermarking is the art of embedding secret messages in multimedia contents in order to protect their intellectual property. While the watermarking of image, audio and video is reaching maturity, the watermarking of 3D virtual objects is still a technology in its infancy. In this thesis, we focus on two main issues. The first one is the perception of the distortions caused by the watermarking process or by attacks on the surface of a 3D model. The second one concerns the development of techniques able to retrieve a watermark without the availability of the original data and after common manipulations and attacks. Since imperceptibility is a strong requirement, assessing the visual perception of the distortions that a 3D model undergoes in the watermarking pipeline is a key issue. In this thesis, we propose an image-based metric that relies on the comparison of 2D views with a Mutual Information criterion. A psychovisual experiment has validated the results of this metric for the most common watermarking attacks. The other issue this thesis deals with is the blind and robust watermarking of 3D shapes. In this context, three different watermarking schemes are proposed. These schemes differ by the classes of 3D watermarking attacks they are able to resist to. The first scheme is based on the extension of spectral decomposition to 3D models. This approach leads to robustness against imperceptible geometric deformations. The weakness of this technique is mainly related to resampling or cropping attacks. The second scheme extends the first to resampling by making use of the automatic multiscale detection of robust umbilical points. The third scheme then addresses the cropping attack by detecting robust prong feature points to locally embed a watermark in the spatial domain.
8

The Construction and Application of Hybrid Factor Model

Tao, Yun-jhen 28 July 2010 (has links)
A Multifactor model is used to explain asset return and risk and its explanatory power depends on common factors that the model uses. Researchers strive to find reasonable factors to enhance multifactor model¡¦s efficiency. However, there are still some unknown factors to be discovered. Miller (2006) presents a general concept and structure of hybrid factor model. The study follows the idea of Miller (2006) and aims to build a complete flow of constructing hybrid factor model that is based on fundamental factor model and statistical factor models. We also apply the hybrid factor model to the Taiwan stock market. We assume that a fundamental factor model is already developed and therefore this study focuses on building the second stage, statistical factor model. Principal Component Analysis is used to form statistical factor and spectral decomposition is used to prepare data for principal component analysis. Those methods are applied to stocks on the Taiwan Stock Exchange in the period of January 1, 2000 to December 31, 2009. This study presents a complete construction flow of hybrid factor models and further confirms that a hybrid factor model is able to find missing factors in a developing market such as Taiwan¡¦s stock market. The study also discovers that the missing factors might be market factor and extensive electronic industry factor.
9

Sparsity and Group Sparsity Constrained Inversion for Spectral Decomposition of Seismic Data

Bonar, Christopher David Unknown Date
No description available.
10

New tools for unsupervised learning

Xiao, Ying 12 January 2015 (has links)
In an unsupervised learning problem, one is given an unlabelled dataset and hopes to find some hidden structure; the prototypical example is clustering similar data. Such problems often arise in machine learning and statistics, but also in signal processing, theoretical computer science, and any number of quantitative scientific fields. The distinguishing feature of unsupervised learning is that there are no privileged variables or labels which are particularly informative, and thus the greatest challenge is often to differentiate between what is relevant or irrelevant in any particular dataset or problem. In the course of this thesis, we study a number of problems which span the breadth of unsupervised learning. We make progress in Gaussian mixtures, independent component analysis (where we solve the open problem of underdetermined ICA), and we formulate and solve a feature selection/dimension reduction model. Throughout, our goal is to give finite sample complexity bounds for our algorithms -- these are essentially the strongest type of quantitative bound that one can prove for such algorithms. Some of our algorithmic techniques turn out to be very efficient in practice as well. Our major technical tool is tensor spectral decomposition: tensors are generalisations of matrices, and often allow access to the "fine structure" of data. Thus, they are often the right tools for unravelling the hidden structure in an unsupervised learning setting. However, naive generalisations of matrix algorithms to tensors run into NP-hardness results almost immediately, and thus to solve our problems, we are obliged to develop two new tensor decompositions (with robust analyses) from scratch. Both of these decompositions are polynomial time, and can be viewed as efficient generalisations of PCA extended to tensors.

Page generated in 0.0447 seconds