• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 40
  • 18
  • 9
  • 3
  • 3
  • 2
  • 1
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 81
  • 81
  • 25
  • 19
  • 19
  • 18
  • 18
  • 17
  • 17
  • 14
  • 14
  • 12
  • 11
  • 11
  • 11
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
51

XSiena: The Content-Based Publish/Subscribe System

Jerzak, Zbigniew 28 September 2009 (has links)
Just as packet switched networks constituted a major breakthrough in our perception of the information exchange in computer networks so have the decoupling properties of publish/subscribe systems revolutionized the way we look at networking in the context of large scale distributed systems. The decoupling of the components of publish/subscribe systems in time, space and synchronization has created an appealing platform for the asynchronous information exchange among anonymous information producers and consumers. Moreover, the content-based nature of publish/subscribe systems provides a great degree of flexibility and expressiveness as far as construction of data flows is considered. However, a number of challenges and not yet addressed issued still exists in the area of the publish/subscribe systems. One active area of research is directed toward the problem of the efficient content delivery in the content-based publish/subscribe networks. Routing of the information based on the information itself, instead of the explicit source and destination addresses poses challenges as far as efficiency and processing times are concerned. Simultaneously, due to their decoupled nature, publish/subscribe systems introduce new challenges with respect to issues related to dependability and fail-awareness. This thesis seeks to advance the field of research in both directions. First it shows the design and implementation of routing algorithms based on the end-to-end systems design principle. Proposed routing algorithms obsolete the need to perform content-based routing within the publish/subscribe network, pushing this task to the edge of the system. Moreover, this thesis presents a fail-aware approach towards construction of the content-based publish/subscribe system along with its application to the creation of the soft state publish/subscribe system. A soft state publish/subscribe system exposes the self stabilizing behavior as far as transient timing, link and node failures are concerned. The result of this thesis is a family of the XSiena content-based publish/subscribe systems, implementing the proposed concepts and algorithms. The family of the XSiena content-based publish/subscribe systems has been a subject to rigorous evaluation, which confirms the claims made in this thesis.
52

Publish Subscribe on Large-Scale Dynamic Topologies: Routing and Overlay Management

Frey, Davide 18 May 2006 (has links) (PDF)
Content-based publish-subscribe is emerging as a communication paradigm able to meet the demands of highly dynamic distributed applications, such as those made popular by mobile computing and peer-to-peer networks. Nevertheless, the available systems implementing this communication model are still unable to cope efficiently with dynamic changes to the topology of their distributed dispatching infrastructure. This hampers their applicability in the aforementioned scenarios. This thesis addresses this problem and presents a complete approach to the reconfiguration of content-based publish-subscribe systems. In Part I, it proposes a layered architecture for reconfigurable publish-subscribe middleware consisting of an overlay, a routing, and an event-recovery layer. This architecture allows the same routing components to operate in different types of dynamic network environments, by exploiting different underlying overlays. Part II addresses the routing layer with new protocols to manage the recon- figuration of the routing information enabling the correct delivery of events to subscribers. When the overlay changes as a result of nodes joining or leaving the network or as a result of mobility, this information is updated so that routing can adapt to the new environment. Our protocols manage to achieve this with as little overhead as possible. Part III addresses the overlay layer and proposes two novel approaches for building and maintaining a connected topology in highly dynamic network sce- narios. The protocols we present achieve this goal, while managing node degree and keeping reconfigurations localized when possible. These properties allow our overlay managers to be applied not only in the context of publish-subscribe mid- dleware but also as enabling technologies for other communication paradigms like application-level multicast. Finally, the thesis integrates the overlay and routing layers into a single frame- work and evaluates their combined performance both in wired and in wireless scenarios. Results show that the optimizations provided by our routing reconfig- uration protocols allow the middleware to achieve very good performance in such networks. Moreover, they highlight that our overlay layer is able to optimize this performance even further, significantly reducing the network traffic generated by the routing layer. The protocols presented in this thesis are implemented in the REDS middle- ware framework developed at Politecnico di Milano. Their use enables REDS to operate efficiently in dynamic network scenarios ranging from large-scale peer-to- peer to mobile ad hoc networks.
53

PS2DICOM: Explorando o paradigma Publish/Subscribe e a elasticidade em níveis aplicados ao procedimento de Telemedicina

Paim, Euclides Palma 31 October 2017 (has links)
Submitted by JOSIANE SANTOS DE OLIVEIRA (josianeso) on 2018-02-22T12:31:13Z No. of bitstreams: 1 Euclides Palma Paim_.pdf: 2529933 bytes, checksum: 9c867ad7f5950b65e99f49343f096e8e (MD5) / Made available in DSpace on 2018-02-22T12:31:13Z (GMT). No. of bitstreams: 1 Euclides Palma Paim_.pdf: 2529933 bytes, checksum: 9c867ad7f5950b65e99f49343f096e8e (MD5) Previous issue date: 2017-10-31 / Nenhuma / Imagens médicas são usadas diariamente para apoio ao diagnóstico em diferentes áreas da Radiologia no mundo todo. Essas imagens seguem uma padronização internacional definida pela ISO 12052, conhecida como padrão DICOM (Digital Imaging and Communications in Medicine). Cada instituição que reivindica conformidade com esta norma, possui seus próprios serviços de armazenamento, sistemas de visualização e processamento específicos para esses dados. No entanto, há uma grande necessidade de que essas imagens sejam analisadas por diferentes especialistas, a fim de que cada caso possa ser discutido de forma ampla, na busca do melhor tratamento para cada patologia. A indisponibilidade de dados em tempo real para a avaliação médica especializada impacta direta e profundamente no sucesso terapêutico. O modelo de computação em nuvem tem as características necessárias para garantir que estas imagens se encontrem ao alcance dos profissionais mais recomendados para cada caso, aptos a oferecer o melhor atendimento. A grande quantidade de recursos disponíveis em nuvem, para lidar com esses dados de forma escalável, facilita a criação de uma infraestrutura para apoio ao diagnóstico à distância através de recursos de Telemedicina. Tomando como base o paradigma computacional Publicar/Assinar, podemos estabelecer comunicação em tempo real para solucionar situações no campo da saúde, como comunicação entre hospitais ou clínicas e entre médicos, enfermeiros e especialistas envolvidos no diagnóstico. Em ambientes clínicos que lidam com transmissão massiva de imagens em alta resolução no padrão DICOM, bem como em ambientes com problemas de desempenho de rede, transmitir essas imagens em tempo hábil, armazenar e disponibilizar de forma segura é um problema sem solução espontânea. Dessa forma esse trabalho propõe uma arquitetura baseada em nuvem computacional, para coletar, comprimir, armazenar e recuperar dados utilizando o paradigma Publicar/Assinar e dois níveis de escalabilidade. O modelo PS2DICOM é estabelecido como um middleware que oferece recursos de infraestrutura na camada IaaS (Infrastructure as a Service), apoiando as tarefas de transmissão e armazenamento de arquivos dentro do padrão DICOM. O sistema oferece compactação dos dados com diferentes intensidades, conforme largura de banda disponível. O modelo PS2DICOM conta ainda com dois níveis de balanceamento de carga e com a elasticidade reativa oferecida pela infraestrutura. A pesquisa contribui ao apresentar uma arquitetura eficaz para otimizar tarefas de rede, capaz de ser adotada como solução ao desenvolver aplicações voltadas para nuvens computacionais aplicadas a saúde em futuras situações reais. A arquitetura foi testada utilizando um protótipo com módulos distintos, desenvolvidos para cada serviço específico oferecido e mostrou-se eficiente como solução para os problemas em questão. Seus detalhes são descritos nos capítulos seguintes, bem como sua implementação, que corrobora a viabilidade do modelo. / Medical images are used daily to support diagnosis in different areas of Radiology throughout the world. These images follow an international standardization defined by ISO 12052, known as DICOM (Digital Imaging and Communications in Medicine) standard. Each institution that claims compliance with this standard has its own storage services, visualization and processing systems specific to that data. However, there is a great need for these images to be analyzed by different specialists, so that each case can be discussed in a broad way, in the search for the best treatment for each pathology. The unavailability of real-time data for specialized medical evaluation has a direct and profound impact on therapeutic success. The cloud computing model has the necessary characteristics to ensure that these images are within the reach of the professionals most recommended for each case, able to offer the best service. The large amount of resources available in the cloud to handle this data in a scalable way facilitates the creation of an infrastructure to support remote diagnosis through Telemedicine resources. Based on the computational paradigm Publish/Subscribe, we can establish real-time communication to solve situations in the field of health, such as communication between hospitals or clinics and between doctors, nurses and experts involved in the diagnosis. In clinical environments that deal with massive transmission of high resolution images in the DICOM standard, as well as in environments with network performance problems, transmitting these images in a timely manner, storing and making them available securely is a problem without a spontaneous solution. In this way, this work proposes a computational cloud-based architecture to collect, compress, store and retrieve data using the Publish/Subscribe paradigm and two levels of scalability. The PS2DICOM model is established as a middleware that provides infrastructure resources in the IaaS (Infrastructure as a Service) layer, supporting the tasks of transmitting and storing files within the DICOM standard. The system offers data compression with different intensities, depending on available bandwidth. The PS2DICOM model also has two levels of load balancing and the reactive elasticity offered by the infrastructure. The research contributes to presenting an efficient architecture to optimize network tasks, capable of being adopted as a solution when developing applications focused on computational clouds applied to health in future real situations. The architecture was tested using a prototype with distinct modules, developed for each specific service offered and proved to be efficient as a solution to the problems in question. Its details are described in the following chapters, as well as its implementation, which corroborates the viability of the model.
54

Hur datakommunikationssäkerheten påverkas vid införandet av en meddelandeförmedlare

Gaupp, Erik, Jonsson, Jan January 2008 (has links)
<p>Validerat; 20101217 (root)</p>
55

The Impress Context Store: A Coordination Framework for Context-Aware Systems

Li, Herman Hon Yu January 2006 (has links)
The dream of weaving technology into our everyday fabric of life is recently being made possible by advances in ubiquitous computing and sensor technologies. Countless sensors of various sizes have made their way into everyday commercial applications. Many projects aim to explore new ways to utilize these new technologies to aid and interact with the general population. Context-aware systems use available context information to assist users automatically, without explicit user input. By inferring user intent and configuring the system proactively for each user, context-aware systems are an integral part of achieving user-friendly ubiquitous computing environments. <br /><br /> A common issue with building a distributed context-aware system is the need to develop a supporting infrastructure providing features such as storage, distributed messaging, and security, before the real work on processing context information can be done. This thesis proposes a coordination framework that provides an effective common foundation for context-aware systems. The separation between the context-processing logic component and the underlying supporting foundation allows researchers to focus their energy at the context-processing part of the system, instead of spending their time re-inventing the supporting infrastructure. <br /><br /> As part of an ongoing project, Impress, the framework uses the open standard, Jabber, as its communication protocol. The Publish-Subscribe (pubsub) extension to Jabber provides interesting features that match those needed by a context-aware system. The main contribution of this thesis is the design and implementation of a coordination framework, called the Impress Context Store, that provides an effective common foundation for context-aware systems. The separation between the context-processing logic and the underlying supporting foundation allows researchers to focus their energy at the context-processing part of the system, instead of spending their time re-inventing the supporting infrastructure.
56

The Impress Context Store: A Coordination Framework for Context-Aware Systems

Li, Herman Hon Yu January 2006 (has links)
The dream of weaving technology into our everyday fabric of life is recently being made possible by advances in ubiquitous computing and sensor technologies. Countless sensors of various sizes have made their way into everyday commercial applications. Many projects aim to explore new ways to utilize these new technologies to aid and interact with the general population. Context-aware systems use available context information to assist users automatically, without explicit user input. By inferring user intent and configuring the system proactively for each user, context-aware systems are an integral part of achieving user-friendly ubiquitous computing environments. <br /><br /> A common issue with building a distributed context-aware system is the need to develop a supporting infrastructure providing features such as storage, distributed messaging, and security, before the real work on processing context information can be done. This thesis proposes a coordination framework that provides an effective common foundation for context-aware systems. The separation between the context-processing logic component and the underlying supporting foundation allows researchers to focus their energy at the context-processing part of the system, instead of spending their time re-inventing the supporting infrastructure. <br /><br /> As part of an ongoing project, Impress, the framework uses the open standard, Jabber, as its communication protocol. The Publish-Subscribe (pubsub) extension to Jabber provides interesting features that match those needed by a context-aware system. The main contribution of this thesis is the design and implementation of a coordination framework, called the Impress Context Store, that provides an effective common foundation for context-aware systems. The separation between the context-processing logic and the underlying supporting foundation allows researchers to focus their energy at the context-processing part of the system, instead of spending their time re-inventing the supporting infrastructure.
57

Dynamic Differential Data Protection for High-Performance and Pervasive Applications

Widener, Patrick M. (Patrick McCall) 20 July 2005 (has links)
Modern distributed applications are long-lived, are expected to provide flexible and adaptive data services, and must meet the functionality and scalability challenges posed by dynamically changing user communities in heterogeneous execution environments. The practical implications of these requirements are that reconfiguration and upgrades are increasingly necessary, but opportunities to perform such tasks offline are greatly reduced. Developers are responding to this situation by dynamically extending or adjusting application functionality and by tuning application performance, a typical method being the incorporation of client- or context-specific code into applications' execution loops. Our work addresses a basic roadblock in deploying such solutions: the protection of key application components and sensitive data in distributed applications. Our approach, termed Dynamic Differential Data Protection (D3P), provides fine-grain methods for providing component-based protection in distributed applications. Context-sensitive, application-specific security methods are deployed at runtime to enforce restrictions in data access and manipulation. D3P is suitable for low- or zero-downtime environments, since deployments are performed while applications run. D3P is appropriate for high performance environments and for highly scalable applications like publish/subscribe, because it creates native codes via dynamic binary code generation. Finally, due to its integration into middleware, D3P can run across a wide variety of operating system and machine platforms. This dissertation introduces D3P, using sample applications from the high performance and pervasive computing domains to illustrate the problems addressed by our D3P solution. It also describes how D3P can be integrated into modern middleware. We present experimental evaluations which demonstrate the fine-grain nature of D3P, that is, its ability to capture individual end users' or components' needs for data protection, and also describe the performance implications of using D3P in data-intensive applications.
58

Scalable view-based techniques for web data : algorithms and systems

Katsifodimos, Asterios 03 July 2013 (has links) (PDF)
XML was recommended by W3C in 1998 as a markup language to be used by device- and system-independent methods of representing information. XML is nowadays used as a data model for storing and querying large volumes of data in database systems. In spite of significant research and systems development, many performance problems are raised by processing very large amounts of XML data. Materialized views have long been used in databases to speed up queries. Materialized views can be seen as precomputed query results that can be re-used to evaluate (part of) another query, and have been a topic of intensive research, in particular in the context of relational data warehousing. This thesis investigates the applicability of materialized views techniques to optimize the performance of Web data management tools, in particular in distributed settings, considering XML data and queries. We make three contributions.We first consider the problem of choosing the best views to materialize within a given space budget in order to improve the performance of a query workload. Our work is the first to address the view selection problem for a rich subset of XQuery. The challenges we face stem from the expressive power and features of both the query and view languages and from the size of the search space of candidate views to materialize. While the general problem has prohibitive complexity, we propose and study a heuristic algorithm and demonstrate its superior performance compared to the state of the art.Second, we consider the management of large XML corpora in peer-to-peer networks, based on distributed hash tables (or DHTs, in short). We consider a platform leveraging distributed materialized XML views, defined by arbitrary XML queries, filled in with data published anywhere in the network, and exploited to efficiently answer queries issued by any network peer. This thesis has contributed important scalability oriented optimizations, as well as a comprehensive set of experiments deployed in a country-wide WAN. These experiments outgrow by orders of magnitude similar competitor systems in terms of data volumes and data dissemination throughput. Thus, they are the most advanced in understanding the performance behavior of DHT-based XML content management in real settings.Finally, we present a novel approach for scalable content-based publish/subscribe (pub/sub, in short) in the presence of constraints on the available computational resources of data publishers. We achieve scalability by off-loading subscriptions from the publisher, and leveraging view-based query rewriting to feed these subscriptions from the data accumulated in others. Our main contribution is a novel algorithm for organizing subscriptions in a multi-level dissemination network in order to serve large numbers of subscriptions, respect capacity constraints, and minimize latency. The efficiency and effectiveness of our algorithm are confirmed through extensive experiments and a large deployment in a WAN.
59

Enabling Internet-Scale Publish/Subscribe In Overlay Networks

Rahimian, Fatemeh January 2011 (has links)
As the amount of data in todays Internet is growing larger, users are exposedto too much information, which becomes increasingly more difficult tocomprehend. Publish/subscribe systems leverage this problem by providingloosely-coupled communications between producers and consumers of data ina network. Data consumers, i.e., subscribers, are provided with a subscriptionmechanism, to express their interests in a subset of data, in order to be notifiedonly when some data that matches their subscription is generated by theproducers, i.e., publishers. Most publish/subscribe systems today, are basedon the client/server architectural model. However, to provide the publish/-subscribe service in large scale, companies either have to invest huge amountof money for over-provisioning the resources, or are prone to frequent servicefailures. Peer-to-peer overlay networks are attractive alternative solutions forbuilding Internet-scale publish/subscribe systems. However, scalability comeswith a cost: a published message often needs to traverse a large number ofuninterested (unsubscribed) nodes before reaching all its subscribers. Werefer to this undesirable traffic, as relay overhead. Without careful considerations,the relay overhead might sharply increase resource consumption for therelay nodes (in terms of bandwidth transmission cost, CPU, etc) and couldultimately lead to rapid deterioration of the system’s performance once therelay nodes start dropping the messages or choose to permanently abandonthe system. To mitigate this problem, some solutions use unbounded numberof connections per node, while some other limit the expressiveness of thesubscription scheme. In this thesis work, we introduce two systems called Vitis and Vinifera, fortopic-based and content-based publish/subscribe models, respectively. Boththese systems are gossip-based and significantly decrease the relay overhead.We utilize novel techniques to cluster together nodes that exhibit similarsubscriptions. In the topic-based model, distinct clusters for each topic areconstructed, while clusters in the content-based model are fuzzy and do nothave explicit boundaries. We augment these clustered overlays by links thatfacilitate routing in the network. We construct a hybrid system by injectingstructure into an otherwise unstructured network. The resulting structuresresemble navigable small-world networks, which spans along clusters of nodesthat have similar subscriptions. The properties of such overlays make theman ideal platform for efficient data dissemination in large-scale systems. Thesystems requires only a bounded node degree and as we show, through simulations,they scale well with the number of nodes and subscriptions and remainefficient under highly complex subscription patterns, high publication rates,and even in the presence of failures in the network. We also compare bothsystems against some state-of-the-art publish/subscribe systems. Our measurementsshow that both Vitis and Vinifera significantly outperform theircounterparts on various subscription and churn scenarios, under both syntheticworkloads and real-world traces. / QC 20111114
60

Efficient Content-Based Publish/Subscribe Systems for Dynamic and Large-Scale Networked Applications

Zhao, Yaxiong January 2012 (has links)
This thesis presents the design and evaluation of content-based publish/subscribe systems for efficient content dissemination and sharing of dynamic and large-scale networked applications. The rapid development of network technologies and the continuous investment in network infrastructure have realized a ubiquitous platform for sharing information. However, there lacks efficient protocol and software that can utilize such resource to support novel networked applications. In this thesis, we explore the possibility of content-based publish/subscribe as an efficient communication substrate for dynamic and large-scale networked applications. Although content-based publish/subscribe has been used extensively in many small-to-medium scale systems, there is no Internet-scale applications that utilize this technology. The research reported in this thesis investigates the technical challenges and their solutions of applying content-based publish/subscribe in various applications in mobile networks and Internet. We apply content-based publish/subscribe in the interest-driven information sharing for smartphone networks. We design efficient approximate content matching algorithms and data structures. We study how to construct optimal overlay publish/subscribe overlay networks. We propose architecture designs that make Internet content-based publish/subscribe robust. We also design a name resolution system that enables content discovery in the Internet. These techniques are evaluated comprehensively in realistic simulation studies, and some of them are further evaluated on PlanetLab testbed with prototype implementations. / Computer and Information Science

Page generated in 0.0485 seconds