Spelling suggestions: "subject:"compressive ensing"" "subject:"compressive densing""
111 |
Dynamics and correlations in sparse signal acquisitionCharles, Adam Shabti 08 June 2015 (has links)
One of the most important parts of engineered and biological systems is the ability to acquire and interpret information from the surrounding world accurately and in time-scales relevant to the tasks critical to system performance. This classical concept of efficient signal acquisition has been a cornerstone of signal processing research, spawning traditional sampling theorems (e.g. Shannon-Nyquist sampling), efficient filter designs (e.g. the Parks-McClellan algorithm), novel VLSI chipsets for embedded systems, and optimal tracking algorithms (e.g. Kalman filtering). Traditional techniques have made minimal assumptions on the actual signals that were being measured and interpreted, essentially only assuming a limited bandwidth. While these assumptions have provided the foundational works in signal processing, recently the ability to collect and analyze large datasets have allowed researchers to see that many important signal classes have much more regularity than having finite bandwidth.
One of the major advances of modern signal processing is to greatly improve on classical signal processing results by leveraging more specific signal statistics. By assuming even very broad classes of signals, signal acquisition and recovery can be greatly improved in regimes where classical techniques are extremely pessimistic. One of the most successful signal assumptions that has gained popularity in recet hears is notion of sparsity. Under the sparsity assumption, the signal is assumed to be composed of a small number of atomic signals from a potentially large dictionary. This limit in the underlying degrees of freedom (the number of atoms used) as opposed to the ambient dimension of the signal has allowed for improved signal acquisition, in particular when the number of measurements is severely limited.
While techniques for leveraging sparsity have been explored extensively in many contexts, typically works in this regime concentrate on exploring static measurement systems which result in static measurements of static signals. Many systems, however, have non-trivial dynamic components, either in the measurement system's operation or in the nature of the signal being observed. Due to the promising prior work leveraging sparsity for signal acquisition and the large number of dynamical systems and signals in many important applications, it is critical to understand whether sparsity assumptions are compatible with dynamical systems. Therefore, this work seeks to understand how dynamics and sparsity can be used jointly in various aspects of signal measurement and inference.
Specifically, this work looks at three different ways that dynamical systems and sparsity assumptions can interact. In terms of measurement systems, we analyze a dynamical neural network that accumulates signal information over time. We prove a series of bounds on the length of the input signal that drives the network that can be recovered from the values at the network nodes~[1--9]. We also analyze sparse signals that are generated via a dynamical system (i.e. a series of correlated, temporally ordered, sparse signals). For this class of signals, we present a series of inference algorithms that leverage both dynamics and sparsity information, improving the potential for signal recovery in a host of applications~[10--19]. As an extension of dynamical filtering, we show how these dynamic filtering ideas can be expanded to the broader class of spatially correlated signals. Specifically, explore how sparsity and spatial correlations can improve inference of material distributions and spectral super-resolution in hyperspectral imagery~[20--25]. Finally, we analyze dynamical systems that perform optimization routines for sparsity-based inference. We analyze a networked system driven by a continuous-time differential equation and show that such a system is capable of recovering a large variety of different sparse signal classes~[26--30].
|
112 |
Near real-time estimation of the seismic source parameters in a compressed domainVera Rodriguez, Ismael A. Unknown Date
No description available.
|
113 |
GPR data processing for reinforced concrete bridge decksWei, Xiangmin 12 January 2015 (has links)
In this thesis, several aspects of GPR data processing for RC bridge decks are studied. First, autofocusing techniques are proposed to replace the previous expensive and unreliable human visual inspections during the iterative migration process for the estimation of the velocity/dielectric permittivity distribution from GPR data. Second, F-K filtering with dip relaxation is proposed for interference removal that is important for both imaging and the performance of post-processing techniques including autofocusing techniques and CS-based migration studied in this thesis. The targeted interferes here are direct waves and cross rebar reflections. The introduced dip relaxation is for accommodating surface roughness and medium inhomogeneity. Third, the newly developed CS-based migration is modified and evaluated on GPR data from RC bridge decks. A more accurate model by accounting for impulse waveform distortion that leads to less modeling errors is proposed. The impact of the selection of the regularization parameter on the comparative amplitude reservation and the imaging performance is also investigated, and an approach to preserve the comparative amplitude information while still maintaining a clear image is proposed. Moreover, the potential of initially sampling the time-spatial data with uniform sampling rates lower than that required by traditional migration methods is evaluated.
|
114 |
Feature detection algorithms in computed imagesGurbuz, Ali Cafer 07 July 2008 (has links)
The problem of sensing a medium by several sensors and retrieving
interesting features is a very general one. The basic framework of the
problem is generally the same for applications from MRI,
tomography, Radar SAR imaging to subsurface imaging, even though the
data acquisition processes, sensing geometries and sensed properties are
different. In this thesis we introduced a new perspective to the
problem of remote sensing and information retrieval by studying the
problem of subsurface imaging using GPR and seismic sensors.
We have shown that if the sensed medium is sparse in some domain then it can be imaged using many fewer measurements than required by the standard methods. This leads to much lower data acquisition times and better images representing the medium. We have used the ideas from Compressive Sensing, which show that a small number of random measurements about a signal is sufficient to completely characterize it, if the signal is sparse or compressible in some domain. Although we have applied our ideas to the subsurface imaging problem, our results are general and can be extended to other remote sensing applications.
A second objective in remote sensing is information retrieval
which involves searching for important features in the computed image of
the medium. In this thesis we focus on detecting buried structures like
pipes, and tunnels in computed GPR or seismic images. The problem of
finding these structures in high clutter and noise conditions, and
finding them faster than the standard shape detecting methods like the
Hough transform is analyzed.
One of the most important contributions of this thesis is, where the
sensing and the information retrieval stages are unified in a single
framework using compressive sensing. Instead of taking lots of standard
measurements to compute the image of the medium and search the
necessary information in the computed image, a much smaller number of
measurements as random projections are taken. The
data acquisition and information retrieval stages are unified by using a
data model dictionary that connects the information to the sensor data.
|
115 |
Channel estimation techniques applied to massive MIMO systems using sparsity and statistics approachesAraújo, Daniel Costa 29 September 2016 (has links)
ARAÚJO, D. C. Channel estimation techniques applied to massive MIMO systems using sparsity and statistics approaches. 2016. 124 f. Tese (Doutorado em Engenharia de Teleinformática)–Centro de
Tecnologia, Universidade Federal do Ceará, Fortaleza, 2016. / Submitted by Renato Vasconcelos (ppgeti@ufc.br) on 2017-06-21T13:52:26Z
No. of bitstreams: 1
2016_tese_dcaraújo.pdf: 1832588 bytes, checksum: a4bb5d44287b92a9321d5fcc3589f22e (MD5) / Approved for entry into archive by Marlene Sousa (mmarlene@ufc.br) on 2017-06-21T16:17:55Z (GMT) No. of bitstreams: 1
2016_tese_dcaraújo.pdf: 1832588 bytes, checksum: a4bb5d44287b92a9321d5fcc3589f22e (MD5) / Made available in DSpace on 2017-06-21T16:17:55Z (GMT). No. of bitstreams: 1
2016_tese_dcaraújo.pdf: 1832588 bytes, checksum: a4bb5d44287b92a9321d5fcc3589f22e (MD5)
Previous issue date: 2016-09-29 / Massive MIMO has the potential of greatly increasing the system spectral efficiency
by employing many individually steerable antenna elements at the base station (BS).
This potential can only be achieved if the BS has sufficient channel state information
(CSI) knowledge. The way of acquiring it depends on the duplexing mode employed
by the communication system. Currently, frequency division duplexing (FDD) is the
most used in the wireless communication system. However, the amount of overhead
necessary to estimate the channel scales with the number of antennas which poses a
big challenge in implementing massive MIMO systems with FDD protocol. To enable
both operating together, this thesis tackles the channel estimation problem by proposing
methods that exploit a compressed version of the massive MIMO channel. There are mainly
two approaches used to achieve such a compression: sparsity and second order statistics. To
derive sparsity-based techniques, this thesis uses a compressive sensing (CS) framework to
extract a sparse-representation of the channel. This is investigated initially in a flat channel
and afterwards in a frequency-selective one. In the former, we show that the Cramer-Rao
lower bound (CRLB) for the problem is a function of pilot sequences that lead to a
Grassmannian matrix. In the frequency-selective case, a novel estimator which combines
CS and tensor analysis is derived. This new method uses the measurements obtained of the
pilot subcarriers to estimate a sparse tensor channel representation. Assuming a Tucker3
model, the proposed solution maps the estimated sparse tensor to a full one which describes
the spatial-frequency channel response. Furthermore, this thesis investigates the problem of
updating the sparse basis that arises when the user is moving. In this study, an algorithm
is proposed to track the arrival and departure directions using very few pilots. Besides
the sparsity-based techniques, this thesis investigates the channel estimation performance
using a statistical approach. In such a case, a new hybrid beamforming (HB) architecture
is proposed to spatially multiplex the pilot sequences and to reduce the overhead. More
specifically, the new solution creates a set of beams that is jointly calculated with the
channel estimator and the pilot power allocation using the minimum mean square error
(MMSE) criterion. We show that this provides enhanced performance for the estimation
process in low signal-noise ratio (SNR) scenarios. / Pesquisas em sistemas MIMO massivo (do inglês multiple-input multiple-output) ganha-
ram muita atenção da comunidade científica devido ao seu potencial em aumentar a
eficiência espectral do sistema comunicações sem-fio utilizando centenas de elementos de
antenas na estação de base (EB). Porém, tal potencial só poderá é obtido se a EB possuir
suficiente informação do estado de canal. A maneira de adquiri-lo depende de como os
recursos de comunicação tempo-frequência são empregados. Atualmente, a solução mais
utilizada em sistemas de comunicação sem fio é a multiplexação por divisão na frequência
(FDD) dos pilotos. Porém, o grande desafio em implementar esse tipo solução é porque
a quantidade de tons pilotos exigidos para estimar o canal aumenta com o número de
antenas. Isso resulta na perda do eficiência espectral prometido pelo sistema massivo.
Esta tese apresenta métodos de estimação de canal que demandam uma quantidade de
tons pilotos reduzida, mas mantendo alta precisão na estimação do canal. Esta redução
de tons pilotos é obtida porque os estimadores propostos exploram a estrutura do canal
para obter uma redução das dimensões do canal. Nesta tese, existem essencialmente duas
abordagens utilizadas para alcançar tal redução de dimensionalidade: uma é através da
esparsidade e a outra através das estatísticas de segunda ordem. Para derivar as soluções
que exploram a esparsidade do canal, o estimador de canal é obtido usando a teoria
de “compressive sensing” (CS) para extrair a representação esparsa do canal. A teoria
é aplicada inicialmente ao problem de estimação de canais seletivos e não-seletivos em
frequência. No primeiro caso, é mostrado que limitante de Cramer-Rao (CRLB) é definido
como uma função das sequências pilotos que geram uma matriz Grassmaniana. No segundo
caso, CS e a análise tensorial são combinado para derivar um novo algoritmo de estimatição
baseado em decomposição tensorial esparsa para canais com seletividade em frequência.
Usando o modelo Tucker3, a solução proposta mapeia o tensor esparso para um tensor
cheio o qual descreve a resposta do canal no espaço e na frequência. Além disso, a tese
investiga a otimização da base de representação esparsa propondo um método para estimar
e corrigir as variações dos ângulos de chegada e de partida, causados pela mobilidade do
usuário. Além das técnicas baseadas em esparsidade, esta tese investida aquelas que usam
o conhecimento estatístico do canal. Neste caso, uma nova arquitetura de beamforming
híbrido é proposta para realizar multiplexação das sequências pilotos. A nova solução
consite em criar um conjunto de feixes, que são calculados conjuntamente com o estimator
de canal e alocação de potência para os pilotos, usand o critério de minimização erro
quadrático médio. É mostrado que esta solução reduz a sequencia pilot e mostra bom
desempenho e cenários de baixa relação sinal ruído (SNR).
|
116 |
Apprentissage de modèles de mélange à large échelle par Sketching / Sketching for large-scale learning of mixture modelsKeriven, Nicolas 12 October 2017 (has links)
Les bases de données modernes sont de très grande taille, parfois divisées et distribuées sur plusieurs lieux de stockage, ou encore sous forme de flux de données : ceci soulève de nouveaux défis majeurs pour les méthodes d’apprentissage statistique. Une des méthodes récentes capable de s’adapter à ces situations consiste à d’abord compresser les données en une structure appelée sketch linéaire, puis ensuite de réaliser la tâche d’apprentissage en utilisant uniquement ce sketch, ce qui est extrêmement rapide si celui-ci est de petite taille. Dans cette thèse, nous définissons une telle méthode pour estimer un modèle de mélange de distributions de probabilités à partir des données, en utilisant uniquement un sketch de celles-ci. Ce sketch est défini en s’inspirant de plusieurs notions venant du domaine des méthodes à noyaux : le plongement par noyau moyen et les approximations aléatoires de noyaux. Défini comme tel, le sketch correspond à des mesures linéaires de la distribution de probabilité sous-jacente aux données. Ainsi nous analysons le problème en utilisant des outils venant du domaine de l’acquisition comprimée, dans lequel un signal est mesuré aléatoirement sans perte d’information, sous certaines conditions. Nous étendons certains résultats de l’acquisition comprimée à la dimension infinie, donnons des conditions génériques garantissant le succès de notre méthode d’estimation de modèles de mélanges, et les appliquons à plusieurs problèmes, dont notamment celui d’estimer des mélanges de distributions stables multivariées, pour lequel il n’existait à ce jour aucun estimateur. Notre analyse est basée sur la construction d’opérateurs de sketch construits aléatoirement, qui satisfont une Propriété d’Isométrie Restreinte dans l’espace de Banach des mesures finies signées avec forte probabilité. Dans une second partie, nous introduisons un algorithme glouton capable heuristiquement d’estimer un modèle de mélange depuis un sketch linéaire. Cet algorithme est appliqué sur données simulées et réelles à trois problèmes : l’estimation de centres significatifs dans les données, pour lequel on constate que la méthode de sketch est significativement plus rapide qu’un algorithme de k-moyennes classique, l’estimation de mélanges de Gaussiennes, pour lequel elle est plus rapide qu’un algorithme d’Espérance-Maximisation, et enfin l’estimation de mélange de distributions stables multivariées, pour lequel il n’existait à ce jour, à notre connaissance, aucun algorithme capable de réaliser une telle tâche. / Learning parameters from voluminous data can be prohibitive in terms of memory and computational requirements. Furthermore, new challenges arise from modern database architectures, such as the requirements for learning methods to be amenable to streaming, parallel and distributed computing. In this context, an increasingly popular approach is to first compress the database into a representation called a linear sketch, that satisfies all the mentioned requirements, then learn the desired information using only this sketch, which can be significantly faster than using the full data if the sketch is small. In this thesis, we introduce a generic methodology to fit a mixture of probability distributions on the data, using only a sketch of the database. The sketch is defined by combining two notions from the reproducing kernel literature, namely kernel mean embedding and Random Features expansions. It is seen to correspond to linear measurements of the underlying probability distribution of the data, and the estimation problem is thus analyzed under the lens of Compressive Sensing (CS), in which a (traditionally finite-dimensional) signal is randomly measured and recovered. We extend CS results to our infinite-dimensional framework, give generic conditions for successful estimation and apply them analysis to many problems, with a focus on mixture models estimation. We base our method on the construction of random sketching operators such that some Restricted Isometry Property (RIP) condition holds in the Banach space of finite signed measures with high probability. In a second part we introduce a flexible heuristic greedy algorithm to estimate mixture models from a sketch. We apply it on synthetic and real data on three problems: the estimation of centroids from a sketch, for which it is seen to be significantly faster than k-means, Gaussian Mixture Model estimation, for which it is more efficient than Expectation-Maximization, and the estimation of mixtures of multivariate stable distributions, for which, to our knowledge, it is the only algorithm capable of performing such a task.
|
117 |
TIME-FREQUENCY ANALYSIS TECHNIQUES FOR NON-STATIONARY SIGNALS USING SPARSITYAMIN, VAISHALI, 0000-0003-0873-3981 January 2022 (has links)
Non-stationary signals, particularly frequency modulated (FM) signals which arecharacterized by their time-varying instantaneous frequencies (IFs), are fundamental
to radar, sonar, radio astronomy, biomedical applications, image processing, speech
processing, and wireless communications. Time-frequency (TF) analyses of such signals
provide two-dimensional mapping of time-domain signals, and thus are regarded
as the most preferred technique for detection, parameter estimation, analysis and
utilization of such signals.
In practice, these signals are often received with compressed measurements as a
result of either missing samples, irregular samplings, or intentional under-sampling of
the signals. These compressed measurements induce undesired noise-like artifacts in
the TF representations (TFRs) of such signals. Compared to random missing data,
burst missing samples present a more realistic, yet a more challenging, scenario for
signal detection and parameter estimation through robust TFRs. In this dissertation,
we investigated the effects of burst missing samples on different joint-variable domain
representations in detail.
Conventional TFRs are not designed to deal with such compressed observations.
On the other hand, sparsity of such non-stationary signals in the TF domain facilitates
utilization of sparse reconstruction-based methods. The limitations of conventional
TF approaches and the sparsity of non-stationary signals in TF domain motivated us
to develop effective TF analysis techniques that enable improved IF estimation of such
signals with high resolution, mitigate undesired effects of cross terms and artifacts
and achieve highly concentrated robust TFRs, which is the goal of this dissertation.
In this dissertation, we developed several TF analysis techniques that achieved
the aforementioned objectives. The developed methods are mainly classified into two
three broad categories: iterative missing data recovery, adaptive local filtering based TF approach, and signal stationarization-based approaches. In the first category,
we recovered the missing data in the instantaneous auto-correlation function (IAF)
domain in conjunction with signal-adaptive TF kernels that are adopted to mitigate
undesired cross-terms and preserve desired auto-terms. In these approaches, we took
advantage of the fact that such non-stationary signals become stationary in the IAF
domain at each time instant. In the second category, we developed a novel adaptive
local filtering-based TF approach that involves local peak detection and filtering of
TFRs within a window of a specified length at each time instant. The threshold for
each local TF segment is adapted based on the local maximum values of the signal
within that segment. This approach offers low-complexity, and is particularly
useful for multi-component signals with distinct amplitude levels. Finally, we developed
knowledge-based TFRs based on signal stationarization and demonstrated
the effectiveness of the proposed TF techniques in high-resolution Doppler analysis
of multipath over-the-horizon radar (OTHR) signals. This is an effective technique
that enables improved target parameter estimation in OTHR operations. However,
due to high proximity of these Doppler signatures in TF domain, their separation
poses a challenging problem. By utilizing signal self-stationarization and ensuring IF
continuity, the developed approaches show excellent performance to handle multiple
signal components with variations in their amplitude levels. / Electrical and Computer Engineering
|
118 |
Robustness, Resilience, and Scalability of State Estimation AlgorithmsShiraz Khan (8782250) 30 November 2023 (has links)
<p dir="ltr">State estimation is a type of an <i>inverse problem</i> in which some amount of observed data needs to be processed using computer algorithms (which are designed using analytical techniques) to infer or reconstruct the underlying model that produced the data. Due to the ubiquity of data and interconnected control systems in the present day, many engineering domains have become replete with inverse problems that can be formulated as state estimation problems. The interconnectedness of these control systems imparts the associated state estimation problems with distinctive structural properties that must be taken into consideration. For instance, the observed data could be high-dimensional and have a dependency structure that is best described by a graph. Furthermore, the control systems of today interface with each other and with the internet, bringing in new possibilities for large-scale collaborative sensor fusion, while also (potentially) introducing new sources of disturbances, faults, and cyberattacks. </p><p dir="ltr">The main thesis of this document is to investigate the unique challenges related to the issues of robustness, resilience (to faults and cyberattacks), and scalability of state estimation algorithms. These correspond to research questions such as, <i>"Does the state estimation algorithm retain its performance when the measurements are perturbed by unknown disturbances or adversarial inputs?"</i> and <i>"Does the algorithm have any bottlenecks that restrict the size/dimension of the problems that it could be applied to?".</i> Most of these research questions are motivated by a singular domain of application: autonomous navigation of unmanned aerial vehicles (UAVs). Nevertheless, the mathematical methods and research philosophy employed herein are quite general, making the results of this document applicable to a variety of engineering tasks, including anomaly detection in time-series data, autonomous remote sensing, traffic monitoring, coordinated motion of dynamical systems, and fault-diagnosis of wireless sensor networks (WSNs), among others.</p>
|
119 |
A Study Of Compressive Sensing For Application To Structural Health MonitoringGanesan, Vaahini 01 January 2014 (has links)
One of the key areas that have attracted attention in the construction industry today is Structural Health Monitoring, more commonly known as SHM. It is a concept developed to monitor the quality and longevity of various engineering structures. The incorporation of such a system would help to continuously track health of the structure, indicate the occurrence/presence of any damage in real time and give us an idea of the number of useful years for the same. Being a recently conceived idea, the state of the art technique in the field is straight forward - populating a given structure with sensors and extracting information from them. In this regard, instrumenting with too many sensors may be inefficient as this could lead to superfluous data that is expensive to capture and process. This research aims to explore an alternate SHM technique that optimizes the data acquisition process by eliminating the amount of redundant data that is sensed and uses this sufficient data to detect and locate the fault present in the structure. Efficient data acquisition requires a mechanism that senses just the necessary amount of data for detection and location of fault. For this reason Compressive Sensing (CS) is explored as a plausible idea. CS claims that signals can be reconstructed from what was previously believed to be incomplete information by Shannon's theorem, taking only a small amount of random and linear non - adaptive measurements. As responses of many physical systems contain a finite basis, CS exploits this feature and determines the sparse solution instead of the traditional least - squares type solution.As a first step, CS is demonstrated by successfully recovering the frequency components of a simple sinusoid. Next, the question of how CS compares with the conventional Fourier transform is analyzed. For this, recovery of temporal frequencies and signal reconstruction is performed using the same number of samples for both the approaches and the errors are compared. On the other hand, the FT error is gradually minimized to match that of CS by increasing the number of regularly placed samples. Once the advantages are established, feasibility of using CS to detect damage in a single degree of freedom system is tested under unforced and forced conditions. In the former scenario, damage is indicated when there is a change in natural frequency of vibration of the system after an impact. In the latter, the system is excited harmonically and damage is detected by a change in amplitude of the system's vibration. As systems in real world applications are predominantly multi-DOF, CS is tested on a 2-DOF system excited with a harmonic forcing. Here again, damage detection is achieved by observing the change in the amplitude of vibration of the system. In order to employ CS for detecting either a change in frequency or amplitude of vibration of a structure subjected to realistic forcing conditions, it would be prudent to explore the reconstruction of a signal which contains multiple frequencies. This is accomplished using CS on a chirp signal. Damage detection is clearly a spatio-temporal problem. Hence it is important to additionally explore the extension of CS to spatial reconstruction. For this reason, mode shape reconstruction of a beam with standard boundary conditions is performed and validated with standard/analytical results from literature. As the final step, the operation deflection shapes (ODS) are reconstructed for a simply supported beam using CS to establish that it is indeed a plausible approach for a less expensive SHM. While experimenting with the idea of spatio-temporal domain, the mode shape as well as the ODS of the given beam are examined under two conditions - undamaged and damaged. Damage in the beam is simulated as a decrease in the stiffness coefficient over a certain number of elements. Although the range of modes to be examined heavily depends on the structure in question, literature suggests that for most practical applications, lower modes are more dominant in indicating damage. For ODS on the other hand, damage is indicated by observing the shift in the recovered spatial frequencies and it is confirmed by the reconstructed response.
|
120 |
On the Use of Physical Basis Functions in a Sparse Expansion for Electromagnetic Scattering SignaturesHalman, Jennifer I. 06 June 2014 (has links)
No description available.
|
Page generated in 0.1009 seconds