• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 11
  • 8
  • 4
  • 2
  • Tagged with
  • 36
  • 36
  • 12
  • 11
  • 10
  • 9
  • 9
  • 7
  • 6
  • 6
  • 6
  • 6
  • 5
  • 5
  • 4
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
21

3D face recognition with wireless transportation

Zou, Le 15 May 2009 (has links)
In this dissertation, we focus on two related parts of a 3D face recognition system with wireless transportation. In the first part, the core components of the system, namely, the feature extraction and classification component, are introduced. In the feature extraction component, range images are taken as inputs and processed in order to extract features. The classification component uses the extracted features as inputs and makes classification decisions based on trained classifiers. In the second part, we consider the wireless transportation problem of range images, which are captured by scattered sensor nodes from target objects and are forwarded to the core components (i.e., feature extraction and classification components) of the face recognition system. Contrary to the conventional definition of being a transducer, a sensor node can be a person, a vehicle, etc. The wireless transportation component not only brings flexibility to the system but also makes the “proactive” face recognition possible. For the feature extraction component, we first introduce the 3D Morphable Model. Then a 3D feature extraction algorithm based on the 3D Morphable Model is presented. The algorithm is insensitive to facial expression. Experimental results show that it can accurately extract features. Following that, we discuss the generic face warping algorithm that can quickly extract features with high accuracy. The proposed algorithm is robust to holes, facial expressions and hair. Furthermore, our experimental results show that the generated features can highly differentiate facial images. For the classification component, a classifier based on Mahalanobis distance is introduced. Based on the classifier, recognition performances of the extracted features are given. The classification results demonstrate the advantage of the features from the generic face warping algorithm. For the wireless transportation of the captured images, we consider the location-based wireless sensor networks (WSN). In order to achieve efficient routing perfor¬mance, a set of distributed stateless routing protocols (PAGER) are proposed for wireless sensor networks. The loop-free and delivery-guaranty properties of the static version (PAGER-S) are proved. Then the performance of PAGER protocols are compared with other well-known routing schemes using network simulator 2 (NS2). Simulation results demonstrate the advantages of PAGER.
22

Automata methods and techniques for graph-structured data

Shoaran, Maryam 23 April 2011 (has links)
Graph-structured data (GSD) is a popular model to represent complex information in a wide variety of applications such as social networks, biological data management, digital libraries, and traffic networks. The flexibility of this model allows the information to evolve and easily integrate with heterogeneous data from many sources. In this dissertation we study three important problems on GSD. A consistent theme of our work is the use of automata methods and techniques to process and reason about GSD. First, we address the problem of answering queries on GSD in a distributed environment. We focus on regular path queries (RPQs) – given by regular expressions matching paths in graph-data. RPQs are the building blocks of almost any mechanism for querying GSD. We present a fault-tolerant, message-efficient, and truly distributed algorithm for answering RPQs. Our algorithm works for the larger class of weighted RPQs on weighted GSDs. Second, we consider the problem of answering RPQs on incomplete GSD, where different data sources are represented by materialized database views. We explore the connection between “certain answers” (CAs) and answers obtained from “view-based rewritings” (VBRs) for RPQs. CAs are answers that can be obtained on each database consistent with the views. Computing all of CAs for RPQs is NP-hard, and one has to resort to an exponential algorithm in the size of the data–view materializations. On the other hand, VBRs are query reformulations in terms of the view definitions. They can be used to obtain query answers in polynomial time in the size of the data. These answers are CAs, but unfortunately for RPQs, not all of the CAs can be obtained in this way. In this work, we show the surprising result that for RPQs under local semantics, using VBRs to answer RPQs gives all the CAs. The importance of this result is that under such semantics, the CAs can be obtained in polynomial time in the size of the data. Third, we focus on XML–an important special case of GSD. The scenario we consider is streaming XML between exchanging parties. The problem we study is flexible validation of streaming XML under the realistic assumption that the schemas of the exchanging parties evolve, and thus diverge from one another. We represent schemas by using Visibly Pushdown Automata (VPAs), which recognize Visibly Pushdown Languages (VPLs). We model evolution for XML by defining formal language operators on VPLs. We show that VPLs are closed under the defined language operators and this enables us to expand the schemas (for XML) in order to account for flexible or constrained evolution. / Graduate
23

Distributed Wireless Networks : Link Scheduling And Application Delay Modelling

Sunny, Albert 05 1900 (has links) (PDF)
We address several problems that arise in a multihop wireless mesh network. First, we study the problem of joint congestion control, routing and MAC layer scheduling. We formulate the problem as an aggregate utility maximization problem and apply duality theory to decompose the problem into two sub-problems, namely, network layer congestion control and routing problem, and MAC layer scheduling problem. Given the link “prices", the source adjusts its rate based on the cost of the least-cost path to the destination, and sends traffic to the destination along the least-cost path, while link scheduling is carried out based on link prices. Optimal link scheduling for a wireless network is known to be NP-hard. We explore the use of a known centralized greedy heuristic, and develop a distributed algorithm that can schedule independent links based on local information. While the link scheduling algorithm above is for a given set of link prices, the solution to our problem depends on the sequence of price vectors generated by the price update algorithm. This leads us to study convergence issues related to the price update algorithm. Next, we develop a practical protocol which maximizes aggregate utility in a wireless mesh network. We simulate our protocol using Qualnet 4.5 and compare the result with a baseline protocol that uses IEEE 802.11 for link scheduling and AODV for routing. Our proposed protocol requires the durations of slots and subslots to be defined. We develop an approach in which given a single cell wireless mesh network using IEEE 802.11 as a reliable message delivery mechanism, one can find upper and lower bounds on the durations of slots. We employ stochastic ordering to compare distributions of random variables and using some properties of stochastic ordering along with the central limit theorem, we devise a way to compute the above mentioned bounds on the durations. In the second part, we shift our focus to model delays incurred by application packets sent over a WLAN. In this section we model the WLAN as a Random Polling System. The packet arrival process at each node i is assumed to be a stationary and independent increment random process with mean ai and second moment a(2)i . The packet lengths at node i are assumed to be i.i.d random variables Pi with finite mean and second moment. Utilizing available results, we obtain expressions for mean packet delay. Extensive simulations are conducted to verify the analytical results.
24

Zertifizierende verteilte Algorithmen

Völlinger, Kim 22 October 2020 (has links)
Eine Herausforderung der Softwareentwicklung ist, die Korrektheit einer Software sicherzustellen. Testen bietet es keine mathematische Korrektheit. Formale Verifikation ist jedoch oft zu aufwändig. Laufzeitverifikation steht zwischen den beiden Methoden. Laufzeitverifikation beantwortet die Frage, ob ein Eingabe-Ausgabe-Paar korrekt ist. Ein zertifizierender Algorithmus überzeugt seinen Nutzer durch ein Korrektheitsargument zur Laufzeit. Dafür berechnet ein zertifizierender Algorithmus für eine Eingabe zusätzlich zur Ausgabe noch einen Zeugen – ein Korrektheitsargument. Jeder zertifizierende Algorithmus besitzt ein Zeugenprädikat: Ist dieses erfüllt für eine Eingabe, eine Ausgabe und einen Zeugen, so ist das Eingabe-Ausgabe-Paar korrekt. Ein simpler Algorithmus, der das Zeugenprädikat für den Nutzer entscheidet, ist ein Checker. Die Korrektheit des Checkers ist folglich notwendig für den Ansatz und die formale Instanzverifikation, bei der wir Checker verifizieren und einen maschinen-geprüften Beweis für die Korrektheit eines Eingabe-Ausgabe-Paars zur Laufzeit gewinnen. Zertifizierende sequentielle Algorithmen sind gut untersucht. Verteilte Algorithmen, die auf verteilten Systemen laufen, unterscheiden sich grundlegend von sequentiellen Algorithmen: die Ausgabe ist über das System verteilt oder der Algorithmus läuft fortwährend. Wir untersuchen zertifizierende verteilte Algorithmen. Unsere Forschungsfrage ist: Wie können wir das Konzept zertifizierender sequentieller Algorithmen so auf verteilte Algorithmen übertragen, dass wir einerseits nah am ursprünglichen Konzept bleiben und andererseits die Gegebenheiten verteilter Systeme berücksichtigen? Wir stellen eine Methode der Übertragung vor. Die beiden Ziele abwägend entwickeln wir eine Klasse zertifizierender verteilter Algorithmen, die verteilte Zeugen berechnen und verteilte Checker besitzen. Wir präsentieren Fallstudien, Entwurfsmuster und ein Framework zur formalen Instanzverifikation. / A major problem in software engineering is to ensure the correctness of software. Testing offers no mathematical correctness. Formal verification is often too costly. Runtime verification stands between the two methods. Runtime verification answers the question whether an input-output pair is correct. A certifying algorithm convinces its user at runtime by offering a correctness argument. For each input, a certifying algorithm computes an output and additionally a witness. Each certifying algorithm has a witness predicate – a predicate with the property: being satisfied for an input, output and witness implies the input-output pair is correct. A simple algorithm deciding the witness predicate for the user is a checker. Hence, the checker’s correctness is crucial to the approach and motivates formal instance verification where we verify checkers and obtain machine-checked proofs for the correctness of an input-output pair at runtime. Certifying sequential algorithms are well-established. Distributed algorithms, designed to run on distributed systems, behave fundamentally different from sequential algorithms: their output is distributed over the system or they even run continuously. We investigate certifying distributed algorithms. Our research question is: How can we transfer the concept of certifying sequential algorithms to distributed algorithms such that we are in line with the original concept but also adapt to the conditions of distributed systems? In this thesis, we present a method to transfer the concept: Weighing up both sometimes conflicting goals, we develop a class of certifying distributed algorithms that compute distributed witnesses and have distributed checkers. We offer case studies, design patterns and a framework for formal instance verification. Additionally, we investigate other methods to transfer the concept of certifying algorithms to distributed algorithms.
25

Non-centralized distributed algorithm to locate nearby servers based on player positions for a MMOG server cluster

Östman, Alexander January 2015 (has links)
In this thesis a non-centralized algorithm is proposed to locate nearby servers based on their players’ positions in a massive multiplayer online game server cluster. The purpose of this is to enable that players can visually see each other even though they are connected to different servers. By utilizing peer to peer connection between the servers the algorithm is tolerant against possible hardware failures. The algorithm simplifies the data sent over the network with a new concave polygon creation algorithm which works in linear execution time, enabling fast computations for real-time games. The algorithm works by finding colliding polygons from other servers and the closest polygons based on distance to find nearby servers which information should be shared with. Those two algorithms at this time work in quadratic execution time which is a point of improvement, which could require the concave polygon to be converted into one or several convex polygons. The algorithm is designed to give the user good access on the amount of network traffic sent over the cluster which gives better control and understanding on how much network traffic that will be sent in the cluster. It shows that the algorithm is dependent on how players in the world are distributed over the servers. By having players nearby each other on the same server improves the result of the algorithm. It is shown that compared to having a centralized server, the network traffic on every single node have reduced network traffic than compared to a centralized server. / In den här uppsatsen presenteras en icke-centraliserad algoritm som hittar närliggande servrar baserat på deras spelares positioner i ett massivt multi-spelare online spel med flera servrar. Syftet är att möjliggöra att spelare från olika servrar kan se varandra visuellt även fast de är uppkopplade till olika servrar. Genom att använda sig av ”peer-to-peer” kommunikation i klustret blir algoritmen tolerant mot hårdvarufel. Algoritmen simplifierar data som skickas genom en ny typ av konkav polygon algoritm vilken fungerar i linjär exekveringstid, vilket möjliggör snabba beräkningar för realtidsspel. Algoritmen fungerar genom att hitta kolliderande polygoner från andra servrar och även de mest närliggande baserat på distans för att lokalisera närliggande servrar att dela information med. De här två algoritmerna arbetar i kvadratisk tid vilket skulle kunna förbättras. Detta kan kräva att konkava polygonen konverteras till en eller flera konvexa polygoner. Algoritmen är designad för att ge användaren bra tillgång till hur mycket nätverkstrafik som bör skickas inom klustret vilket ger en bättre kontroll och förståelse över hur mycket data som kommer att skickas totalt. Det visas att algoritmen är beroende av hur spelarna är distribuerade över servrarna. Genom att ha närliggande spelare i världen på samma server förbättras resultatet av algoritmen. Det visas även att jämfört med en centraliserad server så förbättras nätverkstrafiken på varje enskild nod jämfört med trafiken som mottogs av den centraliserade servern.
26

DECENTRALIZED KEY GENERATION SCHEME FOR CELLULAR-BASED HETEROGENEOUS WIRELESS Ad Hoc NETWORKS

GUPTA, ANANYA 02 October 2006 (has links)
No description available.
27

Navigating Uncertainty: Distributed and Bandit Solutions for Equilibrium Learning in Multiplayer Games

Yuanhanqing Huang (18361527) 15 April 2024 (has links)
<p dir="ltr">In multiplayer games, a collection of self-interested players aims to optimize their individual cost functions in a non-cooperative manner. The cost function of each player depends not only on its own actions but also on the actions of others. In addition, players' actions may also collectively satisfy some global constraints. The study of this problem has grown immensely in the past decades with applications arising in a wide range of societal systems, including strategic behaviors in power markets, traffic assignment of strategic risk-averse users, engagement of multiple humanitarian organizations in disaster relief, etc. Furthermore, with machine learning models playing an increasingly important role in practical applications, the robustness of these models becomes another prominent concern. Investigation into the solutions of multiplayer games and Nash equilibrium problems (NEPs) can advance the algorithm design for fitting these models in the presence of adversarial noises. </p><p dir="ltr">Most of the existing methods for solving multiplayer games assume the presence of a central coordinator, which, unfortunately, is not practical in many scenarios. Moreover, in addition to couplings in the objectives and the global constraints, all too often, the objective functions contain uncertainty in the form of stochastic noises and unknown model parameters. The problem is further complicated by the following considerations: the individual objectives of players may be unavailable or too complex to model; players may exhibit reluctance to disclose their actions; players may experience random delays when receiving feedback regarding their actions. To contend with these issues and uncertainties, in the first half of the thesis, we develop several algorithms based on the theory of operator splitting and stochastic approximation, where the game participants only share their local information and decisions with their trusted neighbors on the network. In the second half of the thesis, we explore the bandit online learning framework as a solution to the challenges, where decisions made by players are updated based solely on the realized objective function values. Our future work will delve into data-driven approaches for learning in multiplayer games and we will explore functional representations of players' decisions, in a departure from the vector form. </p>
28

Algorithmes auto-stabilisants pour la construction de structures couvrantes réparties / Self-Stabilizing Algorithms for Constructing Distributed Spanning Structures

Rivierre, Yvan 12 December 2013 (has links)
Cette thèse s'intéresse à la construction auto-stabilisante de structures couvrantes dans un système réparti. L'auto-stabilisation est un paradigme pour la tolérance aux fautes dans les algorithmes répartis. Plus précisément, elle garantit que le système retrouve un comportement correct en temps fini après avoir été perturbé par des fautes transitoires. Notre modèle de système réparti se base sur des mémoires localement partagées pour la communication, des identifiants uniques pour briser les symétries et un ordonnanceur inéquitable, c'est-à-dire le plus faible des ordonnanceurs. Dans la mesure du possible, nous nous imposons d'utiliser les plus faibles hypothèses, afin d'obtenir les constructions les plus générales de structures couvrantes réparties. Nous présentons quatre algorithmes auto-stabilisants originaux pour le k-partitionnement, la construction d'une (f,g)-alliance et l'indexation. Pour chacun de ces problèmes, nous prouvons la correction de nos solutions. De plus, nous analysons leur complexité en temps et en espace à l'aide de preuves formelles et de simulations. Enfin, pour le problème de (f,g)-alliance, nous prenons en compte la notion de convergence sûre qui vient s'ajouter à celle d'auto-stabilisation. Elle garantit d'abord que le comportement du système assure rapidement un minimum de conditions, puis qu'il continue de converger jusqu'à se conformer à une spécification plus exigeante. / This thesis deals with the self-stabilizing construction of spanning structures over a distributed system. Self-stabilization is a paradigm for fault-tolerance in distributed algorithms. It guarantees that the system eventually satisfies its specification after transient faults hit the system. Our model of distributed system assumes locally shared memories for communicating, unique identifiers for symmetry-breaking, and distributed daemon for execution scheduling, that is, the weakest proper daemon. More generally, we aim for the weakest possible assumptions, such as arbitrary topologies, in order to propose the most versatile constructions of distributed spanning structures. We present four original self-stabilizing algorithms achieving k-clustering, (f,g)-alliance construction, and ranking. For every of these problems, we prove the correctness of our solutions. Moreover, we analyze their time and space complexity using formal proofs and simulations. Finally, for the (f,g)-alliance problem, we consider the notion of safe convergence in addition to self-stabilization. It enforces the system to first quickly satisfy a specification that guarantees a minimum of conditions, and then to converge to a more stringent specification.
29

Distributed Energy-Efficient Solutions for Area Coverage Problems in Wireless Sensor Networks

Vu, Chinh Trung 11 June 2009 (has links)
Wireless sensor networks (WSNs) have recently attracted a great deal of attention due to their numerous attractive applications in many different fields. Sensors and WSNs possess a number of special characteristics that make them very promising in a wide range of applications, but they also put on them lots of constraints that make issues in sensor network particularly challenging. These issues may include topology control, routing, coverage, security, data management and many others. Among them, coverage problem is one of the most fundamental ones for which a WSN has to watch over the environment such as a forest (area coverage) or set of subjects such as collection of precious renaissance paintings (target of point coverage) in order for the network to be able to collect environment parameters, and maybe further monitor the environment. In this dissertation, we highly focus on the area coverage problem. With no assumption of sensors’ locations (i.e., the sensor network is randomly deployed), we only consider distributed and parallel scheduling methods with the ultimate objective of maximizing network lifetime. Additionally, the proposed solutions (including algorithms, a scheme, and a framework) have to be energy-efficient. Generally, we investigate numerous generalizations and variants of the basic coverage problem. Those problems of interest include k-coverage, composite event detection, partial coverage, and coverage for adjustable sensing range network. Various proposed algorithms. In addition, a scheme and a framework are also suggested to solve those problems. The scheme, which is designed for emergency alarming applications, specifies the guidelines for data and communication patterns that significantly reduce the energy consumption and guarantee very low notification delay. For partial coverage problem, we propose a universal framework (consisting of four strategies) which can take almost any complete-coverage algorithm as an input to generate an algorithm for partial coverage. Among the four strategies, two pairs of strategies are trade-off in terms of network lifetime and coverage uniformity. Extensive simulations are conducted to validate the efficiency of each of our proposed solutions.
30

Delay-sensitive Communications Code-Rates, Strategies, and Distributed Control

Parag, Parimal 2011 December 1900 (has links)
An ever increasing demand for instant and reliable information on modern communication networks forces codewords to operate in a non-asymptotic regime. To achieve reliability for imperfect channels in this regime, codewords need to be retransmitted from receiver to the transmit buffer, aided by a fast feedback mechanism. Large occupancy of this buffer results in longer communication delays. Therefore, codewords need to be designed carefully to reduce transmit queue-length and thus the delay experienced in this buffer. We first study the consequences of physical layer decisions on the transmit buffer occupancy. We develop an analytical framework to relate physical layer channel to the transmit buffer occupancy. We compute the optimal code-rate for finite-length codewords operating over a correlated channel, under certain communication service guarantees. We show that channel memory has a significant impact on this optimal code-rate. Next, we study the delay in small ad-hoc networks. In particular, we find out what rates can be supported on a small network, when each flow has a certain end-to-end service guarantee. To this end, service guarantee at each intermediate link is characterized. These results are applied to study the potential benefits of setting up a network suitable for network coding in multicast. In particular, we quantify the gains of network coding over classic routing for service provisioned multicast communication over butterfly networks. In the wireless setting, we study the trade-off between communications gains achieved by network coding and the cost to set-up a network enabling network coding. In particular, we show existence of scenarios where one should not attempt to create a network suitable for coding. Insights obtained from these studies are applied to design a distributed rate control algorithm in a large network. This algorithm maximizes sum-utility of all flows, while satisfying per-flow end-to-end service guarantees. We introduce a notion of effective-capacity per communication link that captures the service requirements of flows sharing this link. Each link maintains a price and effective-capacity, and each flow maintains rate and dissatisfaction. Flows and links update their respective variables locally, and we show that their decisions drive the system to an optimal point. We implemented our algorithm on a network simulator and studied its convergence behavior on few networks of practical interest.

Page generated in 0.0871 seconds