• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 40
  • 18
  • 9
  • 3
  • 3
  • 2
  • 1
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 81
  • 81
  • 25
  • 19
  • 19
  • 18
  • 18
  • 17
  • 17
  • 14
  • 14
  • 12
  • 11
  • 11
  • 11
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
11

Un intergiciel gérant des événements pour permettre l’émergence d’interactions dynamiques et ubiquitaires dans l’Internet des services / Pushing dynamic and ubiquitous event-based interactions in the Internet of services : a middleware for event clouds

Pellegrino, Laurent 03 April 2014 (has links)
Resource Description Framework (RDF) est devenu un modèle de données pertinentafin de décrire et de modéliser les informations qui sont partagées sur le Web.Cependant, fournir une solution permettant de stocker et de récupérer cesdonnées de manière efficace tout en passant à l’échelle reste un défi majeur.Dans le contexte de cette thèse nous proposons un intergiciel dévoué austockage, à la récupération synchrone mais aussi à la dissémination sélectiveet asynchrone en quasi temps réel d'informations RDF dans un environnementcomplètement distribué. L’objectif est de pouvoir tirer parti des informationsdu passé comme de celles filtrées en quasi temps réel. Contrairement à unegrande majorité de solutions existantes, nous avons avons fait le choixd’éviter le hachage pour indexer les données ce qui nous permet de traiter lesrequêtes à intervalles de manière efficace. Le filtrage des informations enquasi temps réel est permis par l’expression des intérêts à l’aide desouscriptions basées sur le contenu des évènements futurs. Nous avons proposédeux algorithmes qui permettent de vérifier la concordance des évènements RDFavec les souscriptions enregistrées. Les deux algorithmes ont été testésexpérimentalement. En sus de la récupération synchrone et de la diffusionasynchrone d’évènements, nous nous sommes intéressés à améliorer la répartitiondes données RDF qui souffrent de dissymétrie. Finalement, nous avons consacréun effort non négligeable à rendre notre intergiciel modulaire. / RDF has become a relevant data model for describing and modeling information on the Web but providing scalable solutions to store and retrieve RDF data in a responsive manner is still challenging. Within the context of this thesis we propose a middleware devoted to storing, retrieving synchronously but also disseminating selectively and asynchronously RDF data in a fully distributed environment. Its purposes is to allow to leverage historical information and filter data near real-time. To this aims we have built our system atop a slightly modified version of a 4-dimensional Content Addressable Network (CAN) overlay network reflecting the structure of an RDF tuple. Unlike many existing solutions we made the choice to avoid hashing for indexing data, thus allowing efficient range queries resolution. Near realtime filtering is enabled by expressing information preferences in advance through content-based subscriptions handled by a publish/subscribe layer designed atop the CAN architecture. We have proposed two algorithms to check RDF data or events satisfaction with subscriptions but also to forward solutions to interested parties. Both algorithms have been experimentally tested for throughput and scalability. Although one performs better than the other, they remain complementary to ensure correctness. Along with information retrieval and dissemination, we have proposed a solution to enhance RDF data distribution on our revised CAN network since RDF information suffers from skewness. Finally, to improve maintainability and reusability some efforts were also dedicated to provide a modular middleware reducing the coupling between its underlying software artifacts.
12

Performance Evaluation of Publish/Subscribe Middleware Architectures / Leistungsuntersuchung von Publish/Subscribe Middleware Architekturen

Henjes, Robert January 2010 (has links) (PDF)
While developing modern applications, it is necessary to ensure an efficient and performant communication between different applications. In current environments, a middleware software is used, which supports the publish/subscribe communication pattern. Using this communication pattern, a publisher sends information encapsulated in messages to the middleware. A subscriber registers its interests at the middleware. The monograph describes three different steps to determine the performance of such a system. In a first step, the message throughput performance of a publish/subscribe in different scenarios is measured using a Java Message Service (JMS) based implementation. In the second step the maximum achievable message throughput is described by adapted models depending on the filter complexity and the replication grade. Using the model, the performance characteristics of a specific system in a given scenario can be determined. These numbers are used for the queuing model described in the third part of the thesis, which supports the dimensioning of a system in realistic scenarios. Additionally, we introduce a method to approximate an M/G/1 system numerically in an efficient way, which can be used for real time analysis to predict the expected performance in a certain scenario. Finally, the analytical model is used to investigate different possibilities to ensure the scalability of the maximum achievable message throughput of the overall system. / Bei der Entwicklung moderner Applikationen ist es notwendig eine effiziente und performante Kommunikation zwischen den einzelnen Anwendungen sicherzustellen. In der Praxis kommt dabei eine Middleware Software zum Einsatz, die das Publish/Subscribe Kommunikationsmuster unterstützt. Dabei senden Publisher Informationen in Form von Nachrichten an die Middleware. Die Subscriber hingegen zeigen durch die Nutzung von Filtern der Middleware an, welche Informationen zugestellt werden sollen. Die Arbeit beschreibt ein dreistufiges Verfahren zur Leistungsbestimmung eines solchen Systems. Zunächst wird durch Messung die Leistung von Publish/Subscribe Systemen in verschiedenen Szenarien untersucht am Beispiel von Java Message Service (JMS) basierten Implementierungen. Danach wird der maximale Nachrichtendurchsatz in Abhängigkeit der Filterkomplexität und des Nachrichtenreplikationsgrades durch einfache Modelle beschrieben. Damit können die Leistungskennwerte für ein System und vorgegebenen Randbedingungen beschrieben werden. Im dritten Teil wird mittels Leistungsbewertung und durch Anwendung eines Warteschlangenmodells die Leistung in praxisnahen Umfeld beschrieben, so dass eine Dimensionierung möglich wird. Zusätzlich wird ein mathematisch, approximatives Verfahren vorgestellt, um ein M/G/1 System numerisch effizient berechnen zu können, was bei der Echtzeitbewertung eines Systems zur Leistungsvorhersage benutzt werden kann. Des Weiteren werden mittels des Modells Möglichkeiten untersucht die Skalierbarkeit des Gesamtsystems in Bezug auf den Nachrichtendurchsatz sicherzustellen.
13

General Boolean Expressions in Publish-Subscribe Systems

Bittner, Sven January 2008 (has links)
The increasing amount of electronically available information in society today is undeniable. Examples include the numbers of general web pages, scientific publications, and items in online auctions. From a user's perspective, this trend will lead to information overflow. Moreover, information publishers are compromised by this situation, as users have greater difficulty in identifying useful information. Publish-subscribe systems can be applied to cope with the reality of information overflow. In these systems, users specify their information interests as subscriptions and, subsequently, only matching information (event messages) is delivered; uninteresting information is filtered out before reaching users. In this dissertation, we consider content-based publish-subscribe systems, a sophisticated example of these systems. They perform the information-filtering task based on the content of provided information. In order to deal with high numbers of subscriptions and frequencies of event messages, publish-subscribe systems are realized as distributed systems. Advertisements---publisher specifications of potential future event messages---are optionally applied in these systems to reduce the internal distribution of subscriptions. Existing work on content-based publish-subscribe concepts mainly focuses on subscriptions and advertisements as pure conjunctive expressions. Therefore, subscriptions or advertisements using operators other than conjunction need to be canonically converted to disjunctive normal form by these systems. Each conjunctive component is then treated as individual subscription or advertisement. Unfortunately, the size of converted expressions is exponential in the worst case. In this dissertation, we show that the direct support of general Boolean subscriptions and advertisements improves the time and space efficiency of general-purpose content-based publish-subscribe systems. For this purpose, we develop suitable approaches for the filtering and routing of general Boolean expressions in these systems. Our approaches represent solutions to exactly those components of content-based publish-subscribe systems that currently restrict subscriptions and advertisements to conjunctive expressions. On the subscription side, we present an effective generic filtering algorithm, and a novel approach to optimize event routing tables, which we call subscription pruning. To support advertisements, we show how to calculate the overlap between subscriptions and advertisements, and introduce the first designated subscription routing optimization, which we refer to as advertisement pruning. We integrate these approaches into our prototype BoP (BOolean Publish-subscribe) which allows for the full support of general Boolean expressions in its filtering and routing components. In the evaluation part of this dissertation, we empirically analyze our prototypical implementation BoP and compare its algorithms to existing conjunctive solutions. We firstly show that our general-purpose Boolean filtering algorithm is more space- and time-efficient than a general-purpose conjunctive filtering algorithm. Secondly, we illustrate the effectiveness of the subscription pruning routing optimization and compare it to the existing covering optimization approach. Finally, we demonstrate the optimization effect of advertisement pruning while maintaining the existing overlapping relationships in the system.
14

Implementation of a Publish/Subscribe Service

Chiou, Min-ling 30 August 2012 (has links)
Over the past few years, a growing attention has been paid to the publish/subscribe (pub/sub) communication paradigm which has become the best model for disseminating information (also called events) through distributed systems on wide-area networks. There are many ways to implement the publish/subscribe system architecture. A common way is publishers and subscribers interact through one or more agents called broker. Broker has to store and management subscriptions, match messages, and efficiently delivery messages to subscribers. It also provides reliability and fault tolerant controlling. Obviously, broker is the most important part of the publish/subscribe system. It is usually required to use a lot of resources such as CPU and memory. In this paper, our publish/subscribe service which implemented by ZeroMQ API could service 10,000 subscribers only cost 3% CPU and 5% memory usage. It is a high performance and low costs publish/subscribe service.
15

RESTful PUBLISH/SUBSCRIBE FRAMEWORK FOR MOBILE DEVICES

2013 November 1900 (has links)
The growing popularity of mobile platforms is changing the Internet user’s computing experience. Current studies suggest that the traditional ubiquitous computing landscape is shifting towards more enhanced and broader mobile computing platform consists of large number of heterogeneous devices. Smartphones and tablets begin to replace the desktop as the primary means of interacting with IT resources. While mobile devices facilitate in consuming web resources in the form of web services, the growing demand for consuming services on mobile device is introducing a complex ecosystem in the mobile environment. This research addresses the communication challenges involved in mobile distributed networks and proposes an event-driven communication approach for information dissemination. This research investigates different communication techniques such as synchronous and asynchronous polling and long-polling, server-side push as mechanisms between client-server interactions and the latest web technologies namely HTML5 standard WebSocket as communication protocol within a publish/subscribe paradigm. Finally, this research introduces and evaluates a framework that is hybrid of REST and event-based publish/subscribe for operating in the mobile environment.
16

Adaptation Techniques for Publish/Subscribe Overlays

Yoon, Young 13 August 2013 (has links)
Publish/Subscribe (in short pub/sub) allows clients that share common interest communicate in an asynchronous and loosely-coupled fashion. This paradigm is adopted by many distributed event-driven applications such as social networking services, distributed business processes and cyber-physical systems. These applications cannot afford to have the underlying pub/sub substrate perform unreliably, permanently fail or behave arbitrarily as it will cause significant disturbance to stably serving many end-users. Therefore, a research effort on making pub/sub systems resilient against various failures to sustain high quality of service to the clients is imperative. In this thesis, we focus on the overlay of pub/sub brokers that are widely adopted as a popular architecture for large-scale pub/sub systems. Broker overlays can suffer from various issues such as degradation of topology quality, brokers causing transient or permanent benign failures and Byzantine brokers behaving arbitrarily. We aim to make novel research contributions by exploring fundamental techniques that can help the broker overlays maintain functional and non-functional requirements even under the presence of the aforementioned failures and necessary administrative updates. We first build a set of overlay adaptation primitives that re-configure topologies such as shifting links and replicating brokers. These primitives are designed to involve a small local group of brokers in the pub/sub overlays so that the disruption during the execution of large-scale and dynamic changes can be controlled in a fined-grained manner. For the problem of degrading topology quality, automated planning systems are developed to find a sequence of adaptations that would cause minimal disruption to running services. Also, our primitives can be executed on demand to quickly fail-over a crashed broker or off-load congested brokers. In addition, these on-demand primitives can be used to form a group of dynamically replicated brokers that enforce a novel safety measure to prevent Byzantine brokers from sabotaging the pub/sub overlays. Our contributions are evaluated with systematic consideration of various trade-offs between functional and non-functional properties.
17

Adaptation Techniques for Publish/Subscribe Overlays

Yoon, Young 13 August 2013 (has links)
Publish/Subscribe (in short pub/sub) allows clients that share common interest communicate in an asynchronous and loosely-coupled fashion. This paradigm is adopted by many distributed event-driven applications such as social networking services, distributed business processes and cyber-physical systems. These applications cannot afford to have the underlying pub/sub substrate perform unreliably, permanently fail or behave arbitrarily as it will cause significant disturbance to stably serving many end-users. Therefore, a research effort on making pub/sub systems resilient against various failures to sustain high quality of service to the clients is imperative. In this thesis, we focus on the overlay of pub/sub brokers that are widely adopted as a popular architecture for large-scale pub/sub systems. Broker overlays can suffer from various issues such as degradation of topology quality, brokers causing transient or permanent benign failures and Byzantine brokers behaving arbitrarily. We aim to make novel research contributions by exploring fundamental techniques that can help the broker overlays maintain functional and non-functional requirements even under the presence of the aforementioned failures and necessary administrative updates. We first build a set of overlay adaptation primitives that re-configure topologies such as shifting links and replicating brokers. These primitives are designed to involve a small local group of brokers in the pub/sub overlays so that the disruption during the execution of large-scale and dynamic changes can be controlled in a fined-grained manner. For the problem of degrading topology quality, automated planning systems are developed to find a sequence of adaptations that would cause minimal disruption to running services. Also, our primitives can be executed on demand to quickly fail-over a crashed broker or off-load congested brokers. In addition, these on-demand primitives can be used to form a group of dynamically replicated brokers that enforce a novel safety measure to prevent Byzantine brokers from sabotaging the pub/sub overlays. Our contributions are evaluated with systematic consideration of various trade-offs between functional and non-functional properties.
18

Distributed resource management with monetary incentives

Tanner, Andreas. Unknown Date (has links) (PDF)
Techn. University, Diss., 2005--Berlin.
19

Kybernetická bezpečnost v systémech typu publish-subscribe / Cyber-security in systems of publish-subscribe type

Detko, Adam January 2020 (has links)
This thesis deals with data distribution service and is divided into three parts. The first part of this thesis focuses on theoretical knowledge of data distribution service. This section gradually introduces the model used in data distribution service, quality of service and the RTPS protocol, which was specially developed for the needs of data distribution service. The most comprehensive part of the theoretical part deals with security in data distribution service. At the end of the first part of this thesis, selected impelementations of data distribution service are introduced. The second part of this thesis deals with practical use of data distribution service, while OpenDDS implementation is chosen for the practical part. This part initially deals with the design of basic communication scenarios, followed by the design of complex communication scenarios. In the next chapters, within the practical part, the thesis is focused on verifyng security of the data distribution service selected implementation, design of security incidents and their subsequent simulation. Content of the final part of this thesis is the introduction of a tool designed for this thesis, which is used to implement selected types of attacks.
20

Traitement continu de requêtes top-k dans les réseaux sociaux / Continuous processing of top-k queries in social networks

Alkhouli, Abdulhafiz 29 September 2017 (has links)
En raison du grand succès des réseaux sociaux, la nature et mode de diffusion del’information sur le Web a changé en faveur de contenus dynamiques diffusés sousforme de flux d’information. Dans le contexte des réseaux sociaux, les utilisateurs peuvent s’abonner à de multiples sources d’information et recevoir continuellement de nouveaux contenus. Or, ce nouveau mode de publication/consommation peut entraîner d’énormes quantités d’information, en surchargeant les utilisateurs. Ainsi,il est essentiel de développer des techniques efficaces de filtrage et de classement qui permettent aux utilisateurs d’être efficacement mis à jour avec le contenu le plus intéressant.Les requêtes top-k sur les flux d’information limitent les résultats au contenu le plus pertinent. Pour améliorer la pertinence des résultats, le modèle de classement des résultats de requêtes devrait tenir compte de divers facteurs de contexte, y compris les facteurs traditionnels basés sur le contenu, les facteurs liés aux utilisateurs et leurs relations (réseau social). Dans le réseau social, le maintien des ensembles de top-k peut être plus difficile car de nombreux événements pourraient changer les messages de top-k tels que le nouveau message, la nouvelle action, le nouvel utilisateur, les modifications de profil, etc. Pour un grand réseau social avec des millions d’utilisateurs et des milliards de messages, le traitement continu des requêtes top-k est l’approche la plus efficace. Cependant, les systèmes actuels pour le traitementcontinu des requêtes top-k échouent lorseque ces systèmes considèrent des modèles de classement riches avec des critères de réseau social. En outre, de tels systèmes ne tiennent pas compte de la diversité des contenus publiés.Dans cette thèse, nous nous concentrons sur le filtrage des flux d’information basé sur le calcul des messages top-k pour chaque utilisateur dans le réseau social. Nous visons à développer un système à large échelle capable d’évaluer efficacement les requêtes top-k continues avec une fonction de classement complexe. Nous proposons l’algorithme SANTA, capable de gérer des fonctions de classement complexes avec des critères sociaux tout en maintenant un traitement continu des requêtes top-k. Nous proposons aussi une variante (SANTA +) qui accélère le traitement d’actions dans les réseaux sociaux. Pour tenire compte de la diversité des contenus publiés, nous proposons l’algorithme DA-SANTA qui étend l’algorithme SANTA pour intégrer la diversité dans le modèle top-k continu tout en maintenant l’efficacité du système. Nos expérimentation sont menées sur des données réelles extraite de Twitter, illustrant les propriétés de nos algorithmes et de montrer leur efficacité. / Information streams provide today a prevalent way of publishing and consuming content on the Web, especially due to the great success of social networks. In the social networks context, users may subscribe to several information sources of interest and continuously receive new published content. But, this new publishing/consumption mode may lead to huge amounts of received information, overwhelming for human processing. Thus, there is a vital need to develop effective filtering and ranking techniques which allow users to efficiently be updated with the most interesting content. Top-k queries over the streams of interest allow limiting results to the most relevant content. To provide a relevant content, the ranking model should consider various context factors including traditional IR factors and social network. In the social network, maintaining top-k sets may be more difficult because many events could produce changes in the top-k sets such as new message, new action, new user, profile changes, etc. For a large social network with millions of users and billionsof messages, the continuous processing of the top-k queries is the most effective approach. However, current systems fail in combining continuous top-k processing with rich scoring models including social network criteria. Moreover, such systems do not consider the diversity of published content.In this thesis, we focus on filtering information streams based on the computation of top-k messages for each user in the social network. We aim to develop a scalable system that be able to efficiently evaluate the continuous top-k queries using the continuous approach with a ranking function including social network criteria. We propose the SANTA algorithm, able to handle scoring functions including content similarity but also social network criteria and events in a continuous processing of top-k queries. We propose a variant (SANTA+) that accelerates the processing of interaction events in social networks. To provide both diverse and relevant messages in top-k sets, we propose the DA-SANTA algorithm which extends the SANTA algorithm to integrate the diversity into the continuous top-k model while maintaining the efficiency of the system. Our experiments are conducted over a real data-set extracted from Twitter, illustrating the properties of our algorithms and demonstrating their efficiency.

Page generated in 0.0895 seconds