• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 9
  • 1
  • 1
  • 1
  • Tagged with
  • 16
  • 16
  • 16
  • 16
  • 11
  • 6
  • 4
  • 4
  • 3
  • 3
  • 3
  • 3
  • 2
  • 2
  • 2
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

Likelihood-Based Modulation Classification for Multiple-Antenna Receivers

Ramezani-Kebrya, Ali 21 September 2012 (has links)
Prior to signal demodulation, blind recognition of the modulation scheme of the received signal is an important task for intelligent radios in various commercial and military applications such as spectrum management, surveillance of broadcasting activities and adaptive transmission. Antenna arrays provide spatial diversity and increase channel capacity. This thesis focuses on the algorithms and performance analysis of the blind modulation classification (MC) for a multiple antenna receiver configuration. For a single-input-multiple-output (SIMO) configuration with unknown channel amplitude, phase, and noise variance, we investigate likelihood-based algorithms for linear digital MC. The existing algorithms are presented and extended to SIMO. Using recently proposed blind estimates of the unknown parameters, a new algorithm is developed. In addition, two upper bounds on the classification performance of MC algorithms are provided. We derive the exact Cramer-Rao Lower Bounds (CRLBs) of joint estimates of the unknown parameters for one- and two-dimensional amplitude modulations. The asymptotic behaviors of the CRLBs are obtained for the high signal-to-noise-ratio (SNR) region. Numerical results demonstrate the accuracy of the CRLB expressions and confirm that the expressions in the literature are special cases of our results. The classification performance of the proposed algorithm is compared with the existing algorithm and upper bounds. It is shown that the proposed algorithm outperforms the existing one significantly with reasonable computational complexity. The proposed algorithm in this thesis can be used in modern intelligent radios equipped with multiple antenna receivers and the provided performance analysis, i.e., the CRLB expressions, can be employed to design practical systems involving estimation of the unknown parameters and is not limited to MC. / Thesis (Master, Electrical & Computer Engineering) -- Queen's University, 2012-09-21 00:51:43.938
2

Estimation and Effects of Imperfect System Parameters on the Performance of Multi-Relay Cooperative Communications Systems

MEHRPOUYAN, HANI 17 September 2012 (has links)
To date the majority of research in the area of cooperative communications focuses on maximizing throughput and reliability while assuming perfect channel state information (CSI) and synchronization. This thesis, seeks to address performance enhancement and system parameter estimation in cooperative networks while relaxing these idealized assumptions. In Chapter 3 the thesis mainly focuses on training-based channel estimation in multi-relay cooperative networks. Channel estimators that are capable of determining the overall channel gains from source to destination antennas are derived. Next, a new low feedback and low complexity scheme is proposed that allows for the coherent combining of signals from multiple relays. Numerical and simulation results show that the combination of the proposed channel estimators and optimization algorithm result in significant performance gains. As communication systems are greatly affected by synchronization parameters, in Chapter 4 the thesis quantitatively analyzes the effects of timing and frequency offset on the performance of communications systems. The modified Cramer-Rao lower bound (MCRLB) undergoing functional transformation, is derived and applied to determine lower bounds on the estimation of signal pulse amplitude and signal-to-noise ratio (SNR) due to timing offset and frequency offset, respectively. In addition, it is shown that estimation of timing and frequency offset can be decoupled in most practical settings. The distributed nature of cooperative relay networks may result in multiple timing and frequency offsets. Chapters 5 and 6 address multiple timing and frequency offset estimation using periodically inserted training sequences in cooperative networks with maximum frequency reuse, i.e., space-division multiple access (SDMA) networks. New closed-form expressions for the Cramer-Rao lower bound (CRLB) for multiple timing and multiple frequency offset estimation for different cooperative protocols are derived. The CRLBs are then applied in a novel way to formulate training sequence design guidelines and determine the effect of network protocol and topology on synchronization parameter estimation. Next, computationally efficient estimators are proposed. Numerical results show that the proposed estimators outperform existing algorithms and reach or approach the CRLB at mid-to-high SNR. When applied to system compensation, simulation results show that application of the proposed estimators allow for synchronized cooperation amongst the nodes within the network. / Thesis (Ph.D, Electrical & Computer Engineering) -- Queen's University, 2010-07-29 16:52:50.272
3

Barometer-Assisted 3D Indoor WiFi Localization for Smart Devices-Map Selection and Performance Evaluation

Ying, Julang 05 May 2016 (has links)
Recently, indoor localization becomes a hot topic no matter in industry or academic field. Smart phones are good candidates for localization since they are carrying various sensors such as GPS, Wi-Fi, accelerometer, barometer and etc, which can be used to estimate the current location. But there are still many challenges for 3D indoor geolocation using smart phones, among which the map selection and 3D performance evaluation problems are the most common and crucial. In the indoor environment, the popular outdoor Google maps cannot be utilized since we need maps showing the layout of every individual floor. Also, layout of different floors differ from one another. Therefore, algorithms are required to detect whether we are inside or outside a building and determine on which floor we are located so that an appropriate map can be selected accordingly. For Wi-Fi based indoor localization, the performance of location estimation is closely related to the algorithms and deployment that we are using. It is difficult to find out a general approach that can be used to evaluate any localization system. On one hand, since the RF signal will suffer extra loss when traveling through the ceilings between floors, its propagation property will be different from the empirical ones and consequently we should design a new propagation model for 3D scenarios. On the other hand, properties of sensors are unique so that corresponding models are required before we analyze the localization scheme. In-depth investigation on the possible hybrid are also needed in case more than one sensor is operated in the localization system. In this thesis, we firstly designed two algorithms to use GPS signal for detecting whether the smart device is operating inside or outside a building, which is called outdoor-indoor transition detection. We also design another algorithm to use barometer data for determining on which floor are we located, which is considered as a multi-floor transition detection. With three scenarios designed inside the Akwater Kent Laboratory building (AK building) at Worcester Polytechnic Institute (WPI), we collected raw data from an Android phone with a version of 4.3 and conducted experimental analysis based on that. An efficient way to quantitatively evaluate the 3D localization systems is using Cramer-Rao Lower Bound (CRLB), which is considered as the lower bound of the estimated error for any localization system. The characteristics of Wi-Fi and barometer signals are explored and proper models are introduced as a foundation. Then we extended the 2D CRLB into a 3D format so that it can fit the our 3D scenarios. A barometer-assisted CRLB is introduced as an improvement for the existing Wi-Fi Receive Signal Strength (RSS)-only scheme and both of the two schemes are compared with the contours in every scenario and the statistical analysis.
4

Performance and Implementation Aspects of Nonlinear Filtering

Hendeby, Gustaf January 2008 (has links)
I många fall är det viktigt att kunna få ut så mycket och så bra information som möjligt ur tillgängliga mätningar. Att utvinna information om till exempel position och hastighet hos ett flygplan kallas för filtrering. I det här fallet är positionen och hastigheten exempel på tillstånd hos flygplanet, som i sin tur är ett system. Ett typiskt exempel på problem av den här typen är olika övervakningssystem, men samma behov blir allt vanligare även i vanliga konsumentprodukter som mobiltelefoner (som talar om var telefonen är), navigationshjälpmedel i bilar och för att placera upplevelseförhöjande grafik i filmer och TV -program. Ett standardverktyg som används för att extrahera den information som behövs är olineär filtrering. Speciellt vanliga är metoderna i positionerings-, navigations- och målföljningstillämpningar. Den här avhandlingen går in på djupet på olika frågeställningar som har med olineär filtrering att göra: * Hur utvärderar man hur bra ett filter eller en detektor fungerar? * Vad skiljer olika metoder åt och vad betyder det för deras egenskaper? * Hur programmerar man de datorer som används för att utvinna informationen? Det mått som oftast används för att tala om hur effektivt ett filter fungerar är RMSE (root mean square error), som i princip är ett mått på hur långt ifrån det korrekta tillståndet man i medel kan förvänta sig att den skattning man får är. En fördel med att använda RMSE som mått är att det begränsas av Cramér-Raos undre gräns (CRLB). Avhandlingen presenterar metoder för att bestämma vilken betydelse olika brusfördelningar har för CRLB. Brus är de störningar och fel som alltid förekommer när man mäter eller försöker beskriva ett beteende, och en brusfördelning är en statistisk beskrivning av hur bruset beter sig. Studien av CRLB leder fram till en analys av intrinsic accuracy (IA), den inneboende noggrannheten i brus. För lineära system får man rättframma resultat som kan användas för att bestämma om de mål som satts upp kan uppnås eller inte. Samma metod kan också användas för att indikera om olineära metoder som partikelfiltret kan förväntas ge bättre resultat än lineära metoder som kalmanfiltret. Motsvarande metoder som är baserade på IA kan även användas för att utvärdera detektionsalgoritmer. Sådana algoritmer används för att upptäcka fel eller förändringar i ett system. När man använder sig av RMSE för att utvärdera filtreringsalgoritmer fångar man upp en aspekt av filtreringsresultatet, men samtidigt finns många andra egenskaper som kan vara intressanta. Simuleringar i avhandlingen visar att även om två olika filtreringsmetoder ger samma prestanda med avseende på RMSE så kan de tillståndsfördelningar de producerar skilja sig väldigt mycket åt beroende på vilket brus det studerade systemet utsätts för. Dessa skillnader kan vara betydelsefulla i vissa fall. Som ett alternativ till RMSE används därför här kullbackdivergensen som tydligt visar på bristerna med att bara förlita sig på RMSE-analyser. Kullbackdivergensen är ett statistiskt mått på hur mycket två fördelningar skiljer sig åt. Två filtreringsalgoritmer har analyserats mer i detalj: det rao-blackwelliserade partikelfiltret (RBPF) och den metod som kallas unscented Kalman filter (UKF). Analysen av RBPF leder fram till ett nytt sätt att presentera algoritmen som gör den lättare att använda i ett datorprogram. Dessutom kan den nya presentationen ge bättre förståelse för hur algoritmen fungerar. I undersökningen av UKF ligger fokus på den underliggande så kallade unscented transformation som används för att beskriva vad som händer med en brusfördelning när man transformerar den, till exempel genom en mätning. Resultatet består av ett antal simuleringsstudier som visar på de olika metodernas beteenden. Ett annat resultat är en jämförelse mellan UT och Gauss approximationsformel av första och andra ordningen. Den här avhandlingen beskriver även en parallell implementation av ett partikelfilter samt ett objektorienterat ramverk för filtrering i programmeringsspråket C ++. Partikelfiltret har implementerats på ett grafikkort. Ett grafikkort är ett exempel på billig hårdvara som sitter i de flesta moderna datorer och mest används för datorspel. Det används därför sällan till sin fulla potential. Ett parallellt partikelfilter, det vill säga ett program som kör flera delar av partikelfiltret samtidigt, öppnar upp för nya tillämpningar där snabbhet och bra prestanda är viktigt. Det objektorienterade ramverket för filtrering uppnår den flexibilitet och prestanda som behövs för storskaliga Monte-Carlo-simuleringar med hjälp av modern mjukvarudesign. Ramverket kan också göra det enklare att gå från en prototyp av ett signalbehandlingssystem till en slutgiltig produkt. / Nonlinear filtering is an important standard tool for information and sensor fusion applications, e.g., localization, navigation, and tracking. It is an essential component in surveillance systems and of increasing importance for standard consumer products, such as cellular phones with localization, car navigation systems, and augmented reality. This thesis addresses several issues related to nonlinear filtering, including performance analysis of filtering and detection, algorithm analysis, and various implementation details. The most commonly used measure of filtering performance is the root mean square error (RMSE), which is bounded from below by the Cramér-Rao lower bound (CRLB). This thesis presents a methodology to determine the effect different noise distributions have on the CRLB. This leads up to an analysis of the intrinsic accuracy (IA), the informativeness of a noise distribution. For linear systems the resulting expressions are direct and can be used to determine whether a problem is feasible or not, and to indicate the efficacy of nonlinear methods such as the particle filter (PF). A similar analysis is used for change detection performance analysis, which once again shows the importance of IA. A problem with the RMSE evaluation is that it captures only one aspect of the resulting estimate and the distribution of the estimates can differ substantially. To solve this problem, the Kullback divergence has been evaluated demonstrating the shortcomings of pure RMSE evaluation. Two estimation algorithms have been analyzed in more detail; the Rao-Blackwellized particle filter (RBPF) by some authors referred to as the marginalized particle filter (MPF) and the unscented Kalman filter (UKF). The RBPF analysis leads to a new way of presenting the algorithm, thereby making it easier to implement. In addition the presentation can possibly give new intuition for the RBPF as being a stochastic Kalman filter bank. In the analysis of the UKF the focus is on the unscented transform (UT). The results include several simulation studies and a comparison with the Gauss approximation of the first and second order in the limit case. This thesis presents an implementation of a parallelized PF and outlines an object-oriented framework for filtering. The PF has been implemented on a graphics processing unit (GPU), i.e., a graphics card. The GPU is a inexpensive parallel computational resource available with most modern computers and is rarely used to its full potential. Being able to implement the PF in parallel makes new applications, where speed and good performance are important, possible. The object-oriented filtering framework provides the flexibility and performance needed for large scale Monte Carlo simulations using modern software design methodology. It can also be used to help to efficiently turn a prototype into a finished product.
5

Analise de tecnicas de localização em redes de sensores sem fio / Analysis of localization techniques in wireless sensor networks

Moreira, Rafael Barbosa 26 February 2007 (has links)
Orientador: Paulo Cardieri / Dissertação (mestrado) - Universidade Estadual de Campinas, Faculdade de Engenharia Eletrica e de Computação / Made available in DSpace on 2018-08-09T19:04:48Z (GMT). No. of bitstreams: 1 Moreira_RafaelBarbosa_M.pdf: 769599 bytes, checksum: 765bba4630a38b38a3832828cf0947b7 (MD5) Previous issue date: 2007 / Resumo: Nesta dissertação, o problema da localização em redes de sensores sem fio é investigado. É apresentada uma análise de desempenho de técnicas de localização por meio de simulação e por meio da avaliação do limite de Cramér-Rao para o erro de localização. Em ambas as formas de análise foram avaliados efeitos de diversos fatores no desempenho, relacionados à topologia da rede e ao ambiente de propagação . Na análise por meio de simulação, foram consideradas as técnicas de localização baseadas em observações de potência do sinal recebido, enquanto que na análise usando o limite de Cramér-Rao, foram analisadas também as técnicas baseadas no tempo de chegada e no ângulo de chegada do sinal recebido. Este trabalho também avaliou os efeitos da polarização das estimativas de distâncias (usadas no processo de localização) no limite inferior de Cramér-Rao. Esta polarização é geralmente desprezada na literatura, o que pode levar a imprecisões no cálculo do limite de Cramér-Rao, em certas condições de propagação. Uma nova expressão para este limite foi derivada para um caso simples de estimador, considerando agora a polarização. Tomando como base o desenvolvimento desta nova expressão, foi derivada também uma nova expressão para o limite inferior de Cramér-Rao considerando os efeitos do desvanecimento lognormal e do desvanecimento Nakagami do canal de propagação / Abstract: This dissertation investigates on the localization problem in wireless sensor networks. A performance analysis of localization techniques through simulations and the Cramér-Rao lower bound is presented. The effects of several parameters on the localization performance are investigated, including network topology and propagation environment. The simulation analysis considered localization techniques based on received signal strength observations, while the Cramér-Rao analysis considered also techniques based on the time of arrival and angle of arrival of the received signal. This work also investigated how the Cramér-Rao limit is affected by the observation bias in localization techniques based on the received signal strength. This bias is usually neglected in the literature, what may lead to imprecisions on the Cramér-Rao limit computation under certain propagation conditions. A new expression for this limit was derived for a simple estimator case, now considering the bias. With the development of this new expression, it was also derived a new expression for the Cramér-Rao lower bound considering the effects of lognormal fading and Nakagami fading on the propagation channel / Mestrado / Telecomunicações e Telemática / Mestre em Engenharia Elétrica
6

Statistical Methods for Image Change Detection with Uncertainty

Lingg, Andrew James January 2012 (has links)
No description available.
7

Statistical Analysis of Geolocation Fundamentals Using Stochastic Geometry

O'Lone, Christopher Edward 22 January 2021 (has links)
The past two decades have seen a surge in the number of applications requiring precise positioning data. Modern cellular networks offer many services based on the user's location, such as emergency services (e.g., E911), and emerging wireless sensor networks are being used in applications spanning environmental monitoring, precision agriculture, warehouse and manufacturing logistics, and traffic monitoring, just to name a few. In these sensor networks in particular, obtaining precise positioning data of the sensors gives vital context to the measurements being reported. While the Global Positioning System (GPS) has traditionally been used to obtain this positioning data, the deployment locations of these cellular and sensor networks in GPS-constrained environments (e.g., cities, indoors, etc.), along with the need for reliable positioning, requires a localization scheme that does not rely solely on GPS. This has lead to localization being performed entirely by the network infrastructure itself, or by the network infrastructure aided, in part, by GPS. In the literature, benchmarking localization performance in these networks has traditionally been done in a deterministic manner. That is, for a fixed setup of anchors (nodes with known location) and a target (a node with unknown location) a commonly used benchmark for localization error, such as the Cramer-Rao lower bound (CRLB), can be calculated for a given localization strategy, e.g., time-of-arrival (TOA), angle-of-arrival (AOA), etc. While this CRLB calculation provides excellent insight into expected localization performance, its traditional treatment as a deterministic value for a specific setup is limited. Rather than trying to gain insight into a specific setup, network designers are more often interested in aggregate localization error statistics within the network as a whole. Questions such as: "What percentage of the time is localization error less than x meters in the network?" are commonplace. In order to answer these types of questions, network designers often turn to simulations; however, these come with many drawbacks, such as lengthy execution times and the inability to provide fundamental insights due to their inherent ``block box'' nature. Thus, this dissertation presents the first analytical solution with which to answer these questions. By leveraging tools from stochastic geometry, anchor positions and potential target positions can be modeled by Poisson point processes (PPPs). This allows for the CRLB of position error to be characterized over all setups of anchor positions and potential target positions realizable within the network. This leads to a distribution of the CRLB, which can completely characterize localization error experienced by a target within the network, and can consequently be used to answer questions regarding network-wide localization performance. The particular CRLB distribution derived in this dissertation is for fourth-generation (4G) and fifth-generation (5G) sub-6GHz networks employing a TOA localization strategy. Recognizing the tremendous potential that stochastic geometry has in gaining new insight into localization, this dissertation continues by further exploring the union of these two fields. First, the concept of localizability, which is the probability that a mobile is able to obtain an unambiguous position estimate, is explored in a 5G, millimeter wave (mm-wave) framework. In this framework, unambiguous single-anchor localization is possible with either a line-of-sight (LOS) path between the anchor and mobile or, if blocked, then via at least two NLOS paths. Thus, for a single anchor-mobile pair in a 5G, mm-wave network, this dissertation derives the mobile's localizability over all environmental realizations this anchor-mobile pair is likely to experience in the network. This is done by: (1) utilizing the Boolean model from stochastic geometry, which statistically characterizes the random positions, sizes, and orientations of reflectors (e.g., buildings) in the environment, (2) considering the availability of first-order (i.e., single-bounce) reflections as well as the LOS path, and (3) considering the possibility that reflectors can either facilitate or block reflections. In addition to the derivation of the mobile's localizability, this analysis also reveals that unambiguous localization, via reflected NLOS signals exclusively, is a relatively small contributor to the mobile's overall localizability. Lastly, using this first-order reflection framework developed under the Boolean model, this dissertation then statistically characterizes the NLOS bias present on range measurements. This NLOS bias is a common phenomenon that arises when trying to measure the distance between two nodes via the time delay of a transmitted signal. If the LOS path is blocked, then the extra distance that the signal must travel to the receiver, in excess of the LOS path, is termed the NLOS bias. Due to the random nature of the propagation environment, the NLOS bias is a random variable, and as such, its distribution is sought. As before, assuming NLOS propagation is due to first-order reflections, and that reflectors can either facilitate or block reflections, the distribution of the path length (i.e., absolute time delay) of the first-arriving multipath component (MPC) is derived. This result is then used to obtain the first NLOS bias distribution in the localization literature that is based on the absolute delay of the first-arriving MPC for outdoor time-of-flight (TOF) range measurements. This distribution is shown to match exceptionally well with commonly assumed gamma and exponential NLOS bias models in the literature, which were only attained previously through heuristic or indirect methods. Finally, the flexibility of this analytical framework is utilized by further deriving the angle-of-arrival (AOA) distribution of the first-arriving MPC at the mobile. This distribution gives novel insight into how environmental obstacles affect the AOA and also represents the first AOA distribution, of any kind, derived under the Boolean model. In summary, this dissertation uses the analytical tools offered by stochastic geometry to gain new insights into localization metrics by performing analyses over the entire ensemble of infrastructure or environmental realizations that a target is likely to experience in a network. / Doctor of Philosophy / The past two decades have seen a surge in the number of applications requiring precise positioning data. Modern cellular networks offer many services based on the user's location, such as emergency services (e.g., E911), and emerging wireless sensor networks are being used in applications spanning environmental monitoring, precision agriculture, warehouse and manufacturing logistics, and traffic monitoring, just to name a few. In these sensor networks in particular, obtaining precise positioning data of the sensors gives vital context to the measurements being reported. While the Global Positioning System (GPS) has traditionally been used to obtain this positioning data, the deployment locations of these cellular and sensor networks in GPS-constrained environments (e.g., cities, indoors, etc.), along with the need for reliable positioning, requires a localization scheme that does not rely solely on GPS. This has lead to localization being performed entirely by the network infrastructure itself, or by the network infrastructure aided, in part, by GPS. When speaking in terms of localization, the network infrastructure consists of what are called anchors, which are simply nodes (points) with a known location. These can be base stations, WiFi access points, or designated sensor nodes, depending on the network. In trying to determine the position of a target (i.e., a user, or a mobile), various measurements can be made between this target and the anchor nodes in close proximity. These measurements are typically distance (range) measurements or angle (bearing) measurements. Localization algorithms then process these measurements to obtain an estimate of the target position. The performance of a given localization algorithm (i.e., estimator) is typically evaluated by examining the distance, in meters, between the position estimates it produces vs. the actual (true) target position. This is called the positioning error of the estimator. There are various benchmarks that bound the best (lowest) error that these algorithms can hope to achieve; however, these benchmarks depend on the particular setup of anchors and the target. The benchmark of localization error considered in this dissertation is the Cramer-Rao lower bound (CRLB). To determine how this benchmark of localization error behaves over the entire network, all of the various setups of anchors and the target that would arise in the network must be considered. Thus, this dissertation uses a field of statistics called stochastic geometry} to model all of these random placements of anchors and the target, which represent all the setups that can be experienced in the network. Under this model, the probability distribution of this localization error benchmark across the entirety of the network is then derived. This distribution allows network designers to examine localization performance in the network as a whole, rather than just for a specific setup, and allows one to obtain answers to questions such as: "What percentage of the time is localization error less than x meters in the network?" Next, this dissertation examines a concept called localizability, which is the probability that a target can obtain a unique position estimate. Oftentimes localization algorithms can produce position estimates that congregate around different potential target positions, and thus, it is important to know when algorithms will produce estimates that congregate around a unique (single) potential target position; hence the importance of localizability. In fifth generation (5G), millimeter wave (mm-wave) networks, only one anchor is needed to produce a unique target position estimate if the line-of-sight (LOS) path between the anchor and the target is unimpeded. If the LOS path is impeded, then a unique target position can still be obtained if two or more non-line-of-sight (NLOS) paths are available. Thus, over all possible environmental realizations likely to be experienced in the network by this single anchor-mobile pair, this dissertation derives the mobile's localizability, or in this case, the probability the LOS path or at least two NLOS paths are available. This is done by utilizing another analytical tool from stochastic geometry known as the Boolean model, which statistically characterizes the random positions, sizes, and orientations of reflectors (e.g., buildings) in the environment. Under this model, considering the availability of first-order (i.e., single-bounce) reflections as well as the LOS path, and considering the possibility that reflectors can either facilitate or block reflections, the mobile's localizability is derived. This result reveals the roles that the LOS path and the NLOS paths play in obtaining a unique position estimate of the target. Using this first-order reflection framework developed under the Boolean model, this dissertation then statistically characterizes the NLOS bias present on range measurements. This NLOS bias is a common phenomenon that arises when trying to measure the distance between two nodes via the time-of-flight (TOF) of a transmitted signal. If the LOS path is blocked, then the extra distance that the signal must travel to the receiver, in excess of the LOS path, is termed the NLOS bias. As before, assuming NLOS propagation is due to first-order reflections and that reflectors can either facilitate or block reflections, the distribution of the path length (i.e., absolute time delay) of the first-arriving multipath component (MPC) (or first-arriving ``reflection path'') is derived. This result is then used to obtain the first NLOS bias distribution in the localization literature that is based on the absolute delay of the first-arriving MPC for outdoor TOF range measurements. This distribution is shown to match exceptionally well with commonly assumed NLOS bias distributions in the literature, which were only attained previously through heuristic or indirect methods. Finally, the flexibility of this analytical framework is utilized by further deriving angle-of-arrival (AOA) distribution of the first-arriving MPC at the mobile. This distribution yields the probability that, for a specific angle, the first-arriving reflection path arrives at the mobile at this angle. This distribution gives novel insight into how environmental obstacles affect the AOA and also represents the first AOA distribution, of any kind, derived under the Boolean model. In summary, this dissertation uses the analytical tools offered by stochastic geometry to gain new insights into localization metrics by performing analyses over all of the possible infrastructure or environmental realizations that a target is likely to experience in a network.
8

TOA-Based Robust Wireless Geolocation and Cramér-Rao Lower Bound Analysis in Harsh LOS/NLOS Environments

Yin, Feng, Fritsche, Carsten, Gustafsson, Fredrik, Zoubir, Abdelhak M January 2013 (has links)
We consider time-of-arrival based robust geolocation in harsh line-of-sight/non-line-of-sight environments. Herein, we assume the probability density function (PDF) of the measurement error to be completely unknown and develop an iterative algorithm for robust position estimation. The iterative algorithm alternates between a PDF estimation step, which approximates the exact measurement error PDF (albeit unknown) under the current parameter estimate via adaptive kernel density estimation, and a parameter estimation step, which resolves a position estimate from the approximate log-likelihood function via a quasi-Newton method. Unless the convergence condition is satisfied, the resolved position estimate is then used to refine the PDF estimation in the next iteration. We also present the best achievable geolocation accuracy in terms of the Cramér-Rao lower bound. Various simulations have been conducted in both real-world and simulated scenarios. When the number of received range measurements is large, the new proposed position estimator attains the performance of the maximum likelihood estimator (MLE). When the number of range measurements is small, it deviates from the MLE, but still outperforms several salient robust estimators in terms of geolocation accuracy, which comes at the cost of higher computational complexity.
9

Statistical methods for reconstruction of entry, descent, and landing performance with application to vehicle design

Dutta, Soumyo 13 January 2014 (has links)
There is significant uncertainty in our knowledge of the Martian atmosphere and the aerodynamics of the Mars entry, descent, and landing (EDL) systems. These uncertainties result in conservatism in the design of the EDL vehicles leading to higher system masses and a broad range of performance predictions. Data from flight instrumentation onboard Mars EDL systems can be used to quantify these uncertainties, but the existing dataset is sparse and many parameters of interest have not been previously observable. Many past EDL reconstructions neither utilize statistical information about the uncertainty of the measured data nor quantify the uncertainty of the estimated parameters. Statistical estimation methods can blend together disparate data types to improve the reconstruction of parameters of interest for the vehicle. For example, integrating data obtained from aeroshell-mounted pressure transducers, inertial measurement unit, and radar altimeter can improve the estimates of the trajectory, atmospheric profile, and aerodynamic coefficients, while also quantifying the uncertainty in these estimates. These same statistical methods can be leveraged to improve current engineering models in order to reduce conservatism in future EDL vehicle design. The work in this thesis presents a comprehensive methodology for parameter reconstruction and uncertainty quantification while blending dissimilar Mars EDL datasets. Statistical estimation methods applied include the Extended Kalman Filter, Unscented Kalman Filter, and Adaptive Filter. The estimators are applied in a manner in which the observability of the parameters of interest is maximized while using the sparse, disparate EDL dataset. The methodology is validated with simulated data and then applied to estimate the EDL performance of the 2012 Mars Science Laboratory. The reconstruction methodology is also utilized as a tool for improving vehicle design and reducing design conservatism. A novel method of optimizing the design of future EDL atmospheric data systems is presented by leveraging the reconstruction methodology. The methodology identifies important design trends and the point of diminishing returns of atmospheric data sensors that are critical in improving the reconstruction performance for future EDL vehicles. The impact of the estimation methodology on aerodynamic and atmospheric engineering models is also studied and suggestions are made for future EDL instrumentation.
10

An Estimation Technique for Spin Echo Electron Paramagnetic Resonance

Golub, Frank 29 August 2013 (has links)
No description available.

Page generated in 0.4942 seconds