• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 85
  • 20
  • 11
  • 8
  • 2
  • 1
  • Tagged with
  • 147
  • 147
  • 46
  • 40
  • 26
  • 22
  • 21
  • 20
  • 18
  • 18
  • 14
  • 14
  • 13
  • 12
  • 12
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
111

Dynamics and correlations in sparse signal acquisition

Charles, Adam Shabti 08 June 2015 (has links)
One of the most important parts of engineered and biological systems is the ability to acquire and interpret information from the surrounding world accurately and in time-scales relevant to the tasks critical to system performance. This classical concept of efficient signal acquisition has been a cornerstone of signal processing research, spawning traditional sampling theorems (e.g. Shannon-Nyquist sampling), efficient filter designs (e.g. the Parks-McClellan algorithm), novel VLSI chipsets for embedded systems, and optimal tracking algorithms (e.g. Kalman filtering). Traditional techniques have made minimal assumptions on the actual signals that were being measured and interpreted, essentially only assuming a limited bandwidth. While these assumptions have provided the foundational works in signal processing, recently the ability to collect and analyze large datasets have allowed researchers to see that many important signal classes have much more regularity than having finite bandwidth. One of the major advances of modern signal processing is to greatly improve on classical signal processing results by leveraging more specific signal statistics. By assuming even very broad classes of signals, signal acquisition and recovery can be greatly improved in regimes where classical techniques are extremely pessimistic. One of the most successful signal assumptions that has gained popularity in recet hears is notion of sparsity. Under the sparsity assumption, the signal is assumed to be composed of a small number of atomic signals from a potentially large dictionary. This limit in the underlying degrees of freedom (the number of atoms used) as opposed to the ambient dimension of the signal has allowed for improved signal acquisition, in particular when the number of measurements is severely limited. While techniques for leveraging sparsity have been explored extensively in many contexts, typically works in this regime concentrate on exploring static measurement systems which result in static measurements of static signals. Many systems, however, have non-trivial dynamic components, either in the measurement system's operation or in the nature of the signal being observed. Due to the promising prior work leveraging sparsity for signal acquisition and the large number of dynamical systems and signals in many important applications, it is critical to understand whether sparsity assumptions are compatible with dynamical systems. Therefore, this work seeks to understand how dynamics and sparsity can be used jointly in various aspects of signal measurement and inference. Specifically, this work looks at three different ways that dynamical systems and sparsity assumptions can interact. In terms of measurement systems, we analyze a dynamical neural network that accumulates signal information over time. We prove a series of bounds on the length of the input signal that drives the network that can be recovered from the values at the network nodes~[1--9]. We also analyze sparse signals that are generated via a dynamical system (i.e. a series of correlated, temporally ordered, sparse signals). For this class of signals, we present a series of inference algorithms that leverage both dynamics and sparsity information, improving the potential for signal recovery in a host of applications~[10--19]. As an extension of dynamical filtering, we show how these dynamic filtering ideas can be expanded to the broader class of spatially correlated signals. Specifically, explore how sparsity and spatial correlations can improve inference of material distributions and spectral super-resolution in hyperspectral imagery~[20--25]. Finally, we analyze dynamical systems that perform optimization routines for sparsity-based inference. We analyze a networked system driven by a continuous-time differential equation and show that such a system is capable of recovering a large variety of different sparse signal classes~[26--30].
112

Near real-time estimation of the seismic source parameters in a compressed domain

Vera Rodriguez, Ismael A. Unknown Date
No description available.
113

GPR data processing for reinforced concrete bridge decks

Wei, Xiangmin 12 January 2015 (has links)
In this thesis, several aspects of GPR data processing for RC bridge decks are studied. First, autofocusing techniques are proposed to replace the previous expensive and unreliable human visual inspections during the iterative migration process for the estimation of the velocity/dielectric permittivity distribution from GPR data. Second, F-K filtering with dip relaxation is proposed for interference removal that is important for both imaging and the performance of post-processing techniques including autofocusing techniques and CS-based migration studied in this thesis. The targeted interferes here are direct waves and cross rebar reflections. The introduced dip relaxation is for accommodating surface roughness and medium inhomogeneity. Third, the newly developed CS-based migration is modified and evaluated on GPR data from RC bridge decks. A more accurate model by accounting for impulse waveform distortion that leads to less modeling errors is proposed. The impact of the selection of the regularization parameter on the comparative amplitude reservation and the imaging performance is also investigated, and an approach to preserve the comparative amplitude information while still maintaining a clear image is proposed. Moreover, the potential of initially sampling the time-spatial data with uniform sampling rates lower than that required by traditional migration methods is evaluated.
114

Feature detection algorithms in computed images

Gurbuz, Ali Cafer 07 July 2008 (has links)
The problem of sensing a medium by several sensors and retrieving interesting features is a very general one. The basic framework of the problem is generally the same for applications from MRI, tomography, Radar SAR imaging to subsurface imaging, even though the data acquisition processes, sensing geometries and sensed properties are different. In this thesis we introduced a new perspective to the problem of remote sensing and information retrieval by studying the problem of subsurface imaging using GPR and seismic sensors. We have shown that if the sensed medium is sparse in some domain then it can be imaged using many fewer measurements than required by the standard methods. This leads to much lower data acquisition times and better images representing the medium. We have used the ideas from Compressive Sensing, which show that a small number of random measurements about a signal is sufficient to completely characterize it, if the signal is sparse or compressible in some domain. Although we have applied our ideas to the subsurface imaging problem, our results are general and can be extended to other remote sensing applications. A second objective in remote sensing is information retrieval which involves searching for important features in the computed image of the medium. In this thesis we focus on detecting buried structures like pipes, and tunnels in computed GPR or seismic images. The problem of finding these structures in high clutter and noise conditions, and finding them faster than the standard shape detecting methods like the Hough transform is analyzed. One of the most important contributions of this thesis is, where the sensing and the information retrieval stages are unified in a single framework using compressive sensing. Instead of taking lots of standard measurements to compute the image of the medium and search the necessary information in the computed image, a much smaller number of measurements as random projections are taken. The data acquisition and information retrieval stages are unified by using a data model dictionary that connects the information to the sensor data.
115

Channel estimation techniques applied to massive MIMO systems using sparsity and statistics approaches

Araújo, Daniel Costa 29 September 2016 (has links)
ARAÚJO, D. C. Channel estimation techniques applied to massive MIMO systems using sparsity and statistics approaches. 2016. 124 f. Tese (Doutorado em Engenharia de Teleinformática)–Centro de Tecnologia, Universidade Federal do Ceará, Fortaleza, 2016. / Submitted by Renato Vasconcelos (ppgeti@ufc.br) on 2017-06-21T13:52:26Z No. of bitstreams: 1 2016_tese_dcaraújo.pdf: 1832588 bytes, checksum: a4bb5d44287b92a9321d5fcc3589f22e (MD5) / Approved for entry into archive by Marlene Sousa (mmarlene@ufc.br) on 2017-06-21T16:17:55Z (GMT) No. of bitstreams: 1 2016_tese_dcaraújo.pdf: 1832588 bytes, checksum: a4bb5d44287b92a9321d5fcc3589f22e (MD5) / Made available in DSpace on 2017-06-21T16:17:55Z (GMT). No. of bitstreams: 1 2016_tese_dcaraújo.pdf: 1832588 bytes, checksum: a4bb5d44287b92a9321d5fcc3589f22e (MD5) Previous issue date: 2016-09-29 / Massive MIMO has the potential of greatly increasing the system spectral efficiency by employing many individually steerable antenna elements at the base station (BS). This potential can only be achieved if the BS has sufficient channel state information (CSI) knowledge. The way of acquiring it depends on the duplexing mode employed by the communication system. Currently, frequency division duplexing (FDD) is the most used in the wireless communication system. However, the amount of overhead necessary to estimate the channel scales with the number of antennas which poses a big challenge in implementing massive MIMO systems with FDD protocol. To enable both operating together, this thesis tackles the channel estimation problem by proposing methods that exploit a compressed version of the massive MIMO channel. There are mainly two approaches used to achieve such a compression: sparsity and second order statistics. To derive sparsity-based techniques, this thesis uses a compressive sensing (CS) framework to extract a sparse-representation of the channel. This is investigated initially in a flat channel and afterwards in a frequency-selective one. In the former, we show that the Cramer-Rao lower bound (CRLB) for the problem is a function of pilot sequences that lead to a Grassmannian matrix. In the frequency-selective case, a novel estimator which combines CS and tensor analysis is derived. This new method uses the measurements obtained of the pilot subcarriers to estimate a sparse tensor channel representation. Assuming a Tucker3 model, the proposed solution maps the estimated sparse tensor to a full one which describes the spatial-frequency channel response. Furthermore, this thesis investigates the problem of updating the sparse basis that arises when the user is moving. In this study, an algorithm is proposed to track the arrival and departure directions using very few pilots. Besides the sparsity-based techniques, this thesis investigates the channel estimation performance using a statistical approach. In such a case, a new hybrid beamforming (HB) architecture is proposed to spatially multiplex the pilot sequences and to reduce the overhead. More specifically, the new solution creates a set of beams that is jointly calculated with the channel estimator and the pilot power allocation using the minimum mean square error (MMSE) criterion. We show that this provides enhanced performance for the estimation process in low signal-noise ratio (SNR) scenarios. / Pesquisas em sistemas MIMO massivo (do inglês multiple-input multiple-output) ganha- ram muita atenção da comunidade científica devido ao seu potencial em aumentar a eficiência espectral do sistema comunicações sem-fio utilizando centenas de elementos de antenas na estação de base (EB). Porém, tal potencial só poderá é obtido se a EB possuir suficiente informação do estado de canal. A maneira de adquiri-lo depende de como os recursos de comunicação tempo-frequência são empregados. Atualmente, a solução mais utilizada em sistemas de comunicação sem fio é a multiplexação por divisão na frequência (FDD) dos pilotos. Porém, o grande desafio em implementar esse tipo solução é porque a quantidade de tons pilotos exigidos para estimar o canal aumenta com o número de antenas. Isso resulta na perda do eficiência espectral prometido pelo sistema massivo. Esta tese apresenta métodos de estimação de canal que demandam uma quantidade de tons pilotos reduzida, mas mantendo alta precisão na estimação do canal. Esta redução de tons pilotos é obtida porque os estimadores propostos exploram a estrutura do canal para obter uma redução das dimensões do canal. Nesta tese, existem essencialmente duas abordagens utilizadas para alcançar tal redução de dimensionalidade: uma é através da esparsidade e a outra através das estatísticas de segunda ordem. Para derivar as soluções que exploram a esparsidade do canal, o estimador de canal é obtido usando a teoria de “compressive sensing” (CS) para extrair a representação esparsa do canal. A teoria é aplicada inicialmente ao problem de estimação de canais seletivos e não-seletivos em frequência. No primeiro caso, é mostrado que limitante de Cramer-Rao (CRLB) é definido como uma função das sequências pilotos que geram uma matriz Grassmaniana. No segundo caso, CS e a análise tensorial são combinado para derivar um novo algoritmo de estimatição baseado em decomposição tensorial esparsa para canais com seletividade em frequência. Usando o modelo Tucker3, a solução proposta mapeia o tensor esparso para um tensor cheio o qual descreve a resposta do canal no espaço e na frequência. Além disso, a tese investiga a otimização da base de representação esparsa propondo um método para estimar e corrigir as variações dos ângulos de chegada e de partida, causados pela mobilidade do usuário. Além das técnicas baseadas em esparsidade, esta tese investida aquelas que usam o conhecimento estatístico do canal. Neste caso, uma nova arquitetura de beamforming híbrido é proposta para realizar multiplexação das sequências pilotos. A nova solução consite em criar um conjunto de feixes, que são calculados conjuntamente com o estimator de canal e alocação de potência para os pilotos, usand o critério de minimização erro quadrático médio. É mostrado que esta solução reduz a sequencia pilot e mostra bom desempenho e cenários de baixa relação sinal ruído (SNR).
116

Apprentissage de modèles de mélange à large échelle par Sketching / Sketching for large-scale learning of mixture models

Keriven, Nicolas 12 October 2017 (has links)
Les bases de données modernes sont de très grande taille, parfois divisées et distribuées sur plusieurs lieux de stockage, ou encore sous forme de flux de données : ceci soulève de nouveaux défis majeurs pour les méthodes d’apprentissage statistique. Une des méthodes récentes capable de s’adapter à ces situations consiste à d’abord compresser les données en une structure appelée sketch linéaire, puis ensuite de réaliser la tâche d’apprentissage en utilisant uniquement ce sketch, ce qui est extrêmement rapide si celui-ci est de petite taille. Dans cette thèse, nous définissons une telle méthode pour estimer un modèle de mélange de distributions de probabilités à partir des données, en utilisant uniquement un sketch de celles-ci. Ce sketch est défini en s’inspirant de plusieurs notions venant du domaine des méthodes à noyaux : le plongement par noyau moyen et les approximations aléatoires de noyaux. Défini comme tel, le sketch correspond à des mesures linéaires de la distribution de probabilité sous-jacente aux données. Ainsi nous analysons le problème en utilisant des outils venant du domaine de l’acquisition comprimée, dans lequel un signal est mesuré aléatoirement sans perte d’information, sous certaines conditions. Nous étendons certains résultats de l’acquisition comprimée à la dimension infinie, donnons des conditions génériques garantissant le succès de notre méthode d’estimation de modèles de mélanges, et les appliquons à plusieurs problèmes, dont notamment celui d’estimer des mélanges de distributions stables multivariées, pour lequel il n’existait à ce jour aucun estimateur. Notre analyse est basée sur la construction d’opérateurs de sketch construits aléatoirement, qui satisfont une Propriété d’Isométrie Restreinte dans l’espace de Banach des mesures finies signées avec forte probabilité. Dans une second partie, nous introduisons un algorithme glouton capable heuristiquement d’estimer un modèle de mélange depuis un sketch linéaire. Cet algorithme est appliqué sur données simulées et réelles à trois problèmes : l’estimation de centres significatifs dans les données, pour lequel on constate que la méthode de sketch est significativement plus rapide qu’un algorithme de k-moyennes classique, l’estimation de mélanges de Gaussiennes, pour lequel elle est plus rapide qu’un algorithme d’Espérance-Maximisation, et enfin l’estimation de mélange de distributions stables multivariées, pour lequel il n’existait à ce jour, à notre connaissance, aucun algorithme capable de réaliser une telle tâche. / Learning parameters from voluminous data can be prohibitive in terms of memory and computational requirements. Furthermore, new challenges arise from modern database architectures, such as the requirements for learning methods to be amenable to streaming, parallel and distributed computing. In this context, an increasingly popular approach is to first compress the database into a representation called a linear sketch, that satisfies all the mentioned requirements, then learn the desired information using only this sketch, which can be significantly faster than using the full data if the sketch is small. In this thesis, we introduce a generic methodology to fit a mixture of probability distributions on the data, using only a sketch of the database. The sketch is defined by combining two notions from the reproducing kernel literature, namely kernel mean embedding and Random Features expansions. It is seen to correspond to linear measurements of the underlying probability distribution of the data, and the estimation problem is thus analyzed under the lens of Compressive Sensing (CS), in which a (traditionally finite-dimensional) signal is randomly measured and recovered. We extend CS results to our infinite-dimensional framework, give generic conditions for successful estimation and apply them analysis to many problems, with a focus on mixture models estimation. We base our method on the construction of random sketching operators such that some Restricted Isometry Property (RIP) condition holds in the Banach space of finite signed measures with high probability. In a second part we introduce a flexible heuristic greedy algorithm to estimate mixture models from a sketch. We apply it on synthetic and real data on three problems: the estimation of centroids from a sketch, for which it is seen to be significantly faster than k-means, Gaussian Mixture Model estimation, for which it is more efficient than Expectation-Maximization, and the estimation of mixtures of multivariate stable distributions, for which, to our knowledge, it is the only algorithm capable of performing such a task.
117

TIME-FREQUENCY ANALYSIS TECHNIQUES FOR NON-STATIONARY SIGNALS USING SPARSITY

AMIN, VAISHALI, 0000-0003-0873-3981 January 2022 (has links)
Non-stationary signals, particularly frequency modulated (FM) signals which arecharacterized by their time-varying instantaneous frequencies (IFs), are fundamental to radar, sonar, radio astronomy, biomedical applications, image processing, speech processing, and wireless communications. Time-frequency (TF) analyses of such signals provide two-dimensional mapping of time-domain signals, and thus are regarded as the most preferred technique for detection, parameter estimation, analysis and utilization of such signals. In practice, these signals are often received with compressed measurements as a result of either missing samples, irregular samplings, or intentional under-sampling of the signals. These compressed measurements induce undesired noise-like artifacts in the TF representations (TFRs) of such signals. Compared to random missing data, burst missing samples present a more realistic, yet a more challenging, scenario for signal detection and parameter estimation through robust TFRs. In this dissertation, we investigated the effects of burst missing samples on different joint-variable domain representations in detail. Conventional TFRs are not designed to deal with such compressed observations. On the other hand, sparsity of such non-stationary signals in the TF domain facilitates utilization of sparse reconstruction-based methods. The limitations of conventional TF approaches and the sparsity of non-stationary signals in TF domain motivated us to develop effective TF analysis techniques that enable improved IF estimation of such signals with high resolution, mitigate undesired effects of cross terms and artifacts and achieve highly concentrated robust TFRs, which is the goal of this dissertation. In this dissertation, we developed several TF analysis techniques that achieved the aforementioned objectives. The developed methods are mainly classified into two three broad categories: iterative missing data recovery, adaptive local filtering based TF approach, and signal stationarization-based approaches. In the first category, we recovered the missing data in the instantaneous auto-correlation function (IAF) domain in conjunction with signal-adaptive TF kernels that are adopted to mitigate undesired cross-terms and preserve desired auto-terms. In these approaches, we took advantage of the fact that such non-stationary signals become stationary in the IAF domain at each time instant. In the second category, we developed a novel adaptive local filtering-based TF approach that involves local peak detection and filtering of TFRs within a window of a specified length at each time instant. The threshold for each local TF segment is adapted based on the local maximum values of the signal within that segment. This approach offers low-complexity, and is particularly useful for multi-component signals with distinct amplitude levels. Finally, we developed knowledge-based TFRs based on signal stationarization and demonstrated the effectiveness of the proposed TF techniques in high-resolution Doppler analysis of multipath over-the-horizon radar (OTHR) signals. This is an effective technique that enables improved target parameter estimation in OTHR operations. However, due to high proximity of these Doppler signatures in TF domain, their separation poses a challenging problem. By utilizing signal self-stationarization and ensuring IF continuity, the developed approaches show excellent performance to handle multiple signal components with variations in their amplitude levels. / Electrical and Computer Engineering
118

Robustness, Resilience, and Scalability of State Estimation Algorithms

Shiraz Khan (8782250) 30 November 2023 (has links)
<p dir="ltr">State estimation is a type of an <i>inverse problem</i> in which some amount of observed data needs to be processed using computer algorithms (which are designed using analytical techniques) to infer or reconstruct the underlying model that produced the data. Due to the ubiquity of data and interconnected control systems in the present day, many engineering domains have become replete with inverse problems that can be formulated as state estimation problems. The interconnectedness of these control systems imparts the associated state estimation problems with distinctive structural properties that must be taken into consideration. For instance, the observed data could be high-dimensional and have a dependency structure that is best described by a graph. Furthermore, the control systems of today interface with each other and with the internet, bringing in new possibilities for large-scale collaborative sensor fusion, while also (potentially) introducing new sources of disturbances, faults, and cyberattacks. </p><p dir="ltr">The main thesis of this document is to investigate the unique challenges related to the issues of robustness, resilience (to faults and cyberattacks), and scalability of state estimation algorithms. These correspond to research questions such as, <i>"Does the state estimation algorithm retain its performance when the measurements are perturbed by unknown disturbances or adversarial inputs?"</i> and <i>"Does the algorithm have any bottlenecks that restrict the size/dimension of the problems that it could be applied to?".</i> Most of these research questions are motivated by a singular domain of application: autonomous navigation of unmanned aerial vehicles (UAVs). Nevertheless, the mathematical methods and research philosophy employed herein are quite general, making the results of this document applicable to a variety of engineering tasks, including anomaly detection in time-series data, autonomous remote sensing, traffic monitoring, coordinated motion of dynamical systems, and fault-diagnosis of wireless sensor networks (WSNs), among others.</p>
119

A Study Of Compressive Sensing For Application To Structural Health Monitoring

Ganesan, Vaahini 01 January 2014 (has links)
One of the key areas that have attracted attention in the construction industry today is Structural Health Monitoring, more commonly known as SHM. It is a concept developed to monitor the quality and longevity of various engineering structures. The incorporation of such a system would help to continuously track health of the structure, indicate the occurrence/presence of any damage in real time and give us an idea of the number of useful years for the same. Being a recently conceived idea, the state of the art technique in the field is straight forward - populating a given structure with sensors and extracting information from them. In this regard, instrumenting with too many sensors may be inefficient as this could lead to superfluous data that is expensive to capture and process. This research aims to explore an alternate SHM technique that optimizes the data acquisition process by eliminating the amount of redundant data that is sensed and uses this sufficient data to detect and locate the fault present in the structure. Efficient data acquisition requires a mechanism that senses just the necessary amount of data for detection and location of fault. For this reason Compressive Sensing (CS) is explored as a plausible idea. CS claims that signals can be reconstructed from what was previously believed to be incomplete information by Shannon's theorem, taking only a small amount of random and linear non - adaptive measurements. As responses of many physical systems contain a finite basis, CS exploits this feature and determines the sparse solution instead of the traditional least - squares type solution.As a first step, CS is demonstrated by successfully recovering the frequency components of a simple sinusoid. Next, the question of how CS compares with the conventional Fourier transform is analyzed. For this, recovery of temporal frequencies and signal reconstruction is performed using the same number of samples for both the approaches and the errors are compared. On the other hand, the FT error is gradually minimized to match that of CS by increasing the number of regularly placed samples. Once the advantages are established, feasibility of using CS to detect damage in a single degree of freedom system is tested under unforced and forced conditions. In the former scenario, damage is indicated when there is a change in natural frequency of vibration of the system after an impact. In the latter, the system is excited harmonically and damage is detected by a change in amplitude of the system's vibration. As systems in real world applications are predominantly multi-DOF, CS is tested on a 2-DOF system excited with a harmonic forcing. Here again, damage detection is achieved by observing the change in the amplitude of vibration of the system. In order to employ CS for detecting either a change in frequency or amplitude of vibration of a structure subjected to realistic forcing conditions, it would be prudent to explore the reconstruction of a signal which contains multiple frequencies. This is accomplished using CS on a chirp signal. Damage detection is clearly a spatio-temporal problem. Hence it is important to additionally explore the extension of CS to spatial reconstruction. For this reason, mode shape reconstruction of a beam with standard boundary conditions is performed and validated with standard/analytical results from literature. As the final step, the operation deflection shapes (ODS) are reconstructed for a simply supported beam using CS to establish that it is indeed a plausible approach for a less expensive SHM. While experimenting with the idea of spatio-temporal domain, the mode shape as well as the ODS of the given beam are examined under two conditions - undamaged and damaged. Damage in the beam is simulated as a decrease in the stiffness coefficient over a certain number of elements. Although the range of modes to be examined heavily depends on the structure in question, literature suggests that for most practical applications, lower modes are more dominant in indicating damage. For ODS on the other hand, damage is indicated by observing the shift in the recovered spatial frequencies and it is confirmed by the reconstructed response.
120

On the Use of Physical Basis Functions in a Sparse Expansion for Electromagnetic Scattering Signatures

Halman, Jennifer I. 06 June 2014 (has links)
No description available.

Page generated in 0.1088 seconds