• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 198
  • 18
  • 18
  • 15
  • 6
  • 5
  • 2
  • 2
  • 2
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 298
  • 298
  • 277
  • 168
  • 168
  • 103
  • 90
  • 77
  • 66
  • 64
  • 48
  • 46
  • 44
  • 38
  • 32
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
181

Lifenet: a flexible ad hoc networking solution for transient environments

Mehendale, Hrushikesh Sanjay 18 November 2011 (has links)
In the wake of major disasters, the failure of existing communications infrastructure and the subsequent lack of an effective communication solution results in increased risks, inefficiencies, damage and casualties. Currently available options such as satellite communication are expensive and have limited functionality. A robust communication solution should be affordable, easy to deploy, require little infrastructure, consume little power and facilitate Internet access. Researchers have long proposed the use of ad hoc wireless networks for such scenarios. However such networks have so far failed to create any impact, primarily because they are unable to handle network transience and have usability constraints such as static topologies and dependence on specific platforms. LifeNet is a WiFi-based ad hoc data communication solution designed for use in highly transient environments. After presenting the motivation, design principles and key insights from prior literature, the dissertation introduces a new routing metric called Reachability and a new routing protocol based on it, called Flexible Routing. Roughly speaking, reachability measures the end-to-end multi-path probability that a packet transmitted by a source reaches its final destination. Using experimental results, it is shown that even with high transience, the reachability metric - (1) accurately captures the effects of transience (2) provides a compact and eventually consistent global network view at individual nodes, (3) is easy to calculate and maintain and (4) captures availability. Flexible Routing trades throughput for availability and fault-tolerance and ensures successful packet delivery under varying degrees of transience. With the intent of deploying LifeNet on field we have been continuously interacting with field partners, one of which is Tata Institute of Social Sciences India. We have refined LifeNet iteratively refined base on their feedback. I conclude the thesis with lessons learned from our field trips so far and deployment plans for the near future.
182

A framework for promoting interoperability in a global electronic market-space

Pather, Maree 30 June 2005 (has links)
The primary contributions to the area of electronic business integration, propounded by this thesis, are (in no particular order):  A novel examination of global Business-to-Business (B2B) interoperability in terms of a "multiplicity paradox" and of a "global electronic market-space" from a Complex Systems Science perspective.  A framework for an, integrated, global electronic market-space, which is based on a hierarchical, incremental, minimalist-business-pattern approach. A Web Services-SOA forms the basis of application-to-application integration within the framework. The framework is founded in a comprehensive study of existing technologies, standards and models for secure interoperability and the SOA paradigm. The Complex Systems Science concepts of "predictable structure" and "structural complexity" are used consistently throughout the progressive formulation of the framework.  A model for a global message handler (including a standards-based message-format) which obviates the common problems implicit in standard SOAP-RPC. It is formulated around the "standardized, common, abstract application interface" critical success factor, deduced from examining existing models. The model can be used in any collaboration context.  An open standards-based security model for the global message handler. Conceptually, the framework comprises the following:  An interoperable standardized message format: a standardized SOAP-envelope with standardized attachments (8-bit binary MIME-serialized XOP packages).  An interoperable standardized message-delivery infrastructure encompassing an RPC-invoked message-handler - a Web service, operating in synchronous and/or asynchronous mode, which relays attachments to service endpoints.  A business information processing infrastructure comprised of: a standardized generic minimalist-business-pattern (simple buying/selling), comprising global pre-specifications for business processes (for example, placing an order), standardized specific atomic business activities (e.g. completing an order-form), a standardized document-set (including, e.g. an order-form) based on standardized metadata (common nomenclature and common semantics used in XSD's, e.g. the order-form), the standardized corresponding choreography for atomic activities (e.g. acknowledgement of receipt of order-form) and service endpoints (based on standardized programming interfaces and virtual methods with customized implementations). / Theoretical Computing / PHD (INFORMATION SYSTEMS)
183

Analytic models of TCP performance

Kassa, Debassey Fesehaye 10 1900 (has links)
Thesis (MSc)--University of Stellenbosch, 2005. / ENGLISH ABSTRACT: The majority of tra c on the Internet uses the Transmission Control Protocol (TCP) as a transport layer protocol for the end-to-end control of information transfer. Measurement, simulation and analytical models are the techniques and tools that can be used to understand and investigate the Internet and its performance. Measurements can only be used to explore existing network scenario or otherwise become costly and in exible with the growth and complexity of the Internet. Simulation models do not scale with the growth of network capacities and the number of users. Computationally e cient analytical models are therefore important tools for investigating, designing, dimensioning and planning IP (Internet Protocol) networks. Existing analytical models of TCP performance are either too simple to capture the internal dynamics of TCP or are too complex to be used to analyze realistic network topologies with several bottleneck links. The literature shows that the xed point algorithm (FPA) is a very useful way of solving analytical models of Internet performance. This thesis presents fast and accurate analytical models of TCP performance with the FPA used to solve them. Apart from what is observed in experimental literature, no comprehensive proof of the convergence and uniqueness of the FPA is given. In this thesis we show how the FPA of analytical models of reliable Internet protocols such as TCP converges to a unique xed point. The thesis speci es the conditions necessary in order to use the FPA for solving analytical models of reliable Internet protocols. We also develop a general implementation algorithm of the FPA of analytical models of TCP performance for realistic and arbitrary network topologies involving heterogenous TCP connections crossing many bottleneck links. The models presented in this thesis give Internet performance metrics, assuming that only basic network parameters such as the network topology, the number of TCP connections, link capacity, distance between network nodes and router bu er sizes are known. To obtain the performance metrics, TCP and network sub{models are used. A closed network of :=G=1 queues is used to develop each TCP sub-model where each queue represents a state of a TCP connection. An M=M=1=K queue is used for each network sub{model which represents the output interface of an IP router with a bu er capacity of K 􀀀��������1 packets. The two sub-models are iteratively solved. We also give closed form expressions for important TCP performance values and distributions. We show how the geometric, bounded geometric and truncated geometric distributions can be used to model reliable protocols such as TCP. We give models of the congestion window cwnd size distribution by conditioning on the slow start threshold ssthresh distribution and vice-versa. We also present models of the probabilities of TCP timeout and triple duplicate ACK receptions. Numerical results based on comparisons against ns2 simulations show that our models are more accurate, simpler and computationally more e cient than another well known TCP model. Our models can therefore be used to rapidly analyze network topologies with several bottlenecks and obtain detailed performance metrics. / AFRIKAANSE OPSOMMING: Die meerderheid van die verkeer op die Internet gebruik die Transmission Control Protocol (TCP) as `n vervoer laag protokol vir die einde-tot-einde kontrole van inligting oordrag. Meting, simulasie en analitiese modelle is die tegnieke en gereedskap wat gebruik kan word om die Internet te ondersoek en verstaan. Meting kan slegs gebruik word om bestaande netwerke scenarios te verken. Meting is duur en onbuigsaam met die groei en samegesteldheid van die Internet. Simulasie modelle skaal nie met die groei van netwerk kapasiteit en gebruikers nie. Analitiese modelle wat berekening e ektief is is dus nodige gereedskap vir die ondersoek, ontwerp, afmeting en beplanning van IP (Internet Protocol) netwerke. Bestaande analitiese TCP modelle is of te eenvoudig om die interne dinamiek van die TCP saam te vat of hulle is te ingewikkeld om realistiese netwerk topologie met heelwat bottelnek skakels te analiseer. Literatuur toon dat die xed point algorithm (FPA) baie handig is vir die oplos van analitiese modelle van Internet verrigting. In hierdie tesis word vinnige en akkurate analitiese modelle van TCP verrigting opgelos deur FPA weergegee. Buiten wat deur eksperimentele literatuur aangedui word is daar geen omvattende bewyse van die konvergensie en uniekheid van die FPA nie. In hierdie tesis word aangedui hoe die FPA van analitiese modelle van betroubare Internet protokolle soos die TCP konvergeer na `n unieke vaste punt. Hierdie tesis spesi seer die voorwaardes benodig om die FPA te gebruik vir die oplos van analitiese modelle van realistiese Internet protokolle. `n Algemene uitvoer algoritme van die FPA van analitiese modelle van TCP vir realistiese en arbitr^ere netwerk topogra e insluitende heterogene TCP konneksies oor baie bottelnek skakels is ontwikkel. Die model in hierdie tesis gee Internet verrigting metodes met die aanname dat slegs basiese netwerk parameters soos netwerk topologie, die aantal TCP konneksies, die konneksie kapasiteit, afstand tussen netwerk nodusse en die roete bu er grotes bekend is. Om die verrigting metodes te verkry, word TCP en netwerk sub-modelle gebruik. `n Geslote netwerk van :=G=1 rye is gebruik om elke TCP sub-model, waar elke ry 'n toestand van `n TCP konneksie voorstel, te ontwikkel. `n M=M=1=K ry is gebruik vir elke netwerk sub-model wat die uitset koppelvlak van `n IP roetemaker met `n bu er kapasiteit van K 􀀀������� 1 pakkies voorstel. Die twee submodelle word iteratief opgelos. Geslote vorm uitdrukkings vir belangrike TCP verrigting waardes en verspreidings word gegee. Daar word getoon hoe geometriese, begrensde geometriese en geknotte geometriese verspreidings gebruik kan word om betroubare protokolle soos die TCP te modelleer. Modelle van die kongestie venster cwnd grootte verspreiding word gegee deur die kondisionering van die stadige aanvang drempel ssthresh verspreiding en andersom. Modelle van die voorspelling van TCP tyduit en trippel duplikaat ACK resepsie word weergegee. Numeriese resultate gebaseer op vergelykings met ns2 simulasies wys dat ons modelle meer akkuraat, eenvoudiger en berekeningsgewys meer e ektief is as ander wel bekende TCP modelle. Ons modelle kan dus gebruik word vir vinnig analise van netwerk topologie met verskeie bottelnekke en om gedetailleerde verrigting metodes te bekom.
184

The development of a dynamically configured wireless ad-hoc multihop network protocol

Pretorius, Wynand 12 1900 (has links)
Thesis (MScEng (Electrical and Electronic Engineering))--University of Stellenbosch, 2006. / An ad-hoc network encompasses the cooperative engagement of a collection of mobile nodes that are free to move and communicate with each other wirelessly without the required intervention of any centralized access point or existing infrastructure. The advantage of such a network lies in it’s robustness, adaptiveness, the fact that its self-configurable and that it becomes somewhat indestructible due to it’s decentralized nature. But such a network layout simultaneously introduces many complex network management issues which are normally taken care of inherently by a rigid network architecture. The biggest challenge faced by any such protocol is the fact that it needs to be scalable, must maintain a decent stable data throughput, all whilst performing it’s own continuous network management and associated routing algorithms. These mobile nodes need a complex, scalable, compact and essentially realtime algorithm for maintaining an up to date representation of the overall network layout, yet without clogging the system’s communications channels with too much overhead traffic, and drastically lowering the effective data throughput. Since each mobile node only has a limited communications range each node also needs very advanced routing capabilities which will allow it to track who is currently within communications range, and at the same time allow the node to create multihop paths to distant destination nodes, thus connecting nodes which cannot directly communicate. This report follows the development process of both the software needed to successfully conceptualize, simulate and test the protocol, as well as the hardware needed as proof of concept. It highlights and discusses the various design choices / considerations made in development of such a protocol, the strong- and weakpoints of the developed protocol, as well as providing several possibilities to further evolve the developed protocol.
185

QUIC-TCP: validation of QUIC-TCP through network simulations

Unknown Date (has links)
The scalability of QUIC-TCP was examined by expanding previous developmental 11-node, 4-flow topology to over 30 nodes with 11 flows to validate QUIC-TCP for larger networks. The topology was simulated using ns-2 network simulator with the same ns-2 module of FAST-TCP modified to produce QUIC-TCP agent that the original development used. A symmetrical topology and a random topology were examined. Fairness, aggregate throughput and the object of the utility function were used as validation criteria. It was shown through simulation that QUICTCP optimized the utility function and demonstrated a good balance between aggregate throughput and fairness; therefore QUIC-TCP is indeed scalable to larger networks. / Includes bibliography. / Thesis (M.S.)--Florida Atlantic University, 2013.
186

Implementation of an Available Bit Rate Service for Satellite IP Networks using a Performance Enhancing Proxy

Reddy, Pavan K 29 April 2004 (has links)
The transport control protocol (TCP) is one of the most heavily used protocols on the Internet, offering a reliable, connection oriented transport service. However, the quality of service (QoS) provided by the TCP protocol deteriorates when it is used over satellite IP networks. With the increased usage of Internet applications by the military in remote geographical regions, there is an increased need to address some of the shortcomings of TCP performance in satellite IP networks. In this research we describe our efforts at designing and testing a performance enhancing proxy (PEP) that can be used improve the QoS provided by the TCP service in large latency networks. We also show how one can use such a proxy to create a new transport service similar to the Available Bit Rate (ABR) service provided by ATM networks without needing ATM infrastructure, this new service offers a connection oriented, reliable, best effort transport service with minimal queuing delay, jitter and throughput variation.
187

Adaptive Explicit Congestion Notification (AECN) for Heterogeneous Flows

Zheng, Zici 02 May 2001 (has links)
Previous research on ECN and RED usually considered only a limited traffic domain, focusing on networks with a small number of homogeneous flows. The behavior of RED and ECN congestion control mechanisms in TCP network with many competing heterogeneous flows in the bottleneck link, hasn't been sufficiently explored. This thesis first investigates the behavior and performance of RED with ECN congestion control mechanisms with many heterogeneous TCP Reno flows using the network simulation tool, ns-2. By comparing the simulated performance of RED and ECN routers, this study finds that ECN does provide better goodput and fairness than RED for heterogeneous flows. However, when the demand is held constant, the number of flows generating the demand has a negative effect on performance. Meanwhile, the simulations with many flows demonstrate that the bottleneck router's marking probability must be aggressively increased to provide good ECN performance. Based on these simulation results, an Adaptive ECN algorithm (AECN) was studied to further improve the goodput and fairness of ECN. AECN divides all flows competing for a bottleneck into three flow groups, and deploys a different max for each flow group. Meanwhile, AECN also adjusts min for the robust flow group and max to get higher performance when the number of flows grows large. Furthermore, AECN uses mark-front strategy, instead of mark-tail strategy in standard ECN. A series of AECN simulations were run in ns-2. The simulations show clearly that AECN treats each flow fairer than ECN with the two fairness measurements: Jain's fairness index and visual max-min fairness. AECN has fewer packet drops and alleviates the lockout phenomenon and yields higher goodput than ECN.
188

Measurement of Windows Streaming Media

Nichols III, James G 22 April 2004 (has links)
The growth of high speed Internet connections has fueled an increase in the demand for high quality streaming video. In order to satisfy timing constraints, streaming video typically uses UDP as the default network transport protocol. Unfortunately, UDP does not have any end-to-end congestion control mechanisms, and so in the absence of higher layer congestion control can lead to unfairness and possibly congestion collapse. While there has been research done in video measurement and characterization using custom tools, to the best of our knowledge, there have been no measurement studies where the researchers had control over a commercial streaming media server and client, and control of the network conditions and content. A goal of this research is to characterize the bitrate response of Windows Streaming Media in response to network-level metrics such as capacity, loss rate, and round-trip time. We build a streaming media test bed that allows us to systematically vary network and content encoding characteristics. We analyze responsiveness by comparing streaming media flows to TCP-friendly flows under various streaming configurations and network conditions. We find Windows Streaming Media has a prominent buffering phase in which it sends data at a bitrate significantly higher than the steady-state rate. Overall, Windows Streaming Media is responsive to available capacity, but is often unfair to TCP. Knowledge of streaming media's response to congestion encountered in the network is important in building networks that better accommodate their turbulence. The additional characteristics we measure can be combined to guide emulation or simulation configurations and network traffic generators for use in further research.
189

Reducing internet latency for thin-stream applications over reliable transport with active queue management

Grigorescu, Eduard January 2018 (has links)
An increasing number of network applications use reliable transport protocols. Applications with constant data transmission recover from loss without major performance disruption, however, applications that send data sporadically, in small packets, also called thin-streams, experience frequently high latencies due to 'Bufferbloat', that reduce the application performance. Active Queue Management mechanisms were proposed to dynamically manage the queues in routers by dropping packets early and reduce these, hence reducing latency. While their deployment to the internet remains an open issue, the proper investigation into how their functioning mechanism impacts latency is the main focus of this work and research questions have been devised to investigate the AQM impact on latency. A range of AQM mechanisms has been evaluated by the research, exploring performance of the methods for latency sensitive network applications. This has explored new single queue AQM mechanisms such as Controlled Delay (CODEL) and Proportional Integral Enhanced (PIE) and Adaptive RED (ARED). The evaluation has shown great improvements in queuing latency when AQM are used over a range of network scenarios. Scheduling AQM algorithms such as FlowQueue CODEL (FQ-CODEL) isolates traffic and minimises the impact of Bufferbloat on flows. The core components of FQ-CODEL, still widely misunderstood at the time of its inception, have been explained in depth by this study and their contribution to reducing latency have been evaluated. The results show significant reductions in queuing latency for thin streams using FQ-CODEL. When TCP is used for thin streams, high application latencies can arise when there are retransmissions, for example after dropping packets by an AQM mechanism. This delay is a result of TCP's loss-based congestion control mechanism that controls sender transmission rate following packet loss. ECN, a marking sender-side improvement to TCP reduces applicationlayer latency without disrupting the overall network performance. The thesis evaluated the benefit of using ECN using a wide range of experiments. The findings show that FQ-CODEL with ECN provides a substantial reduction of application latency compared to a drop-based AQM. Moreover, this study recommends the combination of FQ-CODEL with other mechanisms, to reduce application latency. Mechanisms such as ABE, have been shown to increase aggregate throughput and reduce application latency for thin-stream applications.
190

Improving throughput and fairness of on-board mobile networks.

Baig, Adeel, Computer Science & Engineering, Faculty of Engineering, UNSW January 2007 (has links)
The Internet Engineering Task Force (IETF) has recently released network mobility standards that allow deployment of TCP/IP networks onboard a vehicle and maintain permanent network connectivity to the Internet via a vehicular mobile router. This recent development opens up new opportunities for providing efficient mobile computing for users on the move, especially for commuters traveling on public transports. Moreover, central and coordinated management of mobility in a single router, rather than by each user device individually, has numerous advantages. In this architecture, however, it becomes challenging to guarantee network performance due to the mobility of the network and inherently vulnerable nature of wireless links. In this thesis, a detailed performance study of onboard networks is conducted. It has been shown that disruptions in the mobile router connectivity can significantly degrade network throughput. Moreover, factors such as the limited wireless bandwidth of the access link, variations in the bandwidth due to technology switching, and the communication diversity of onboard users all contribute to the problem of unfair sharing of wireless bandwidth. By leveraging the fact that all onboard communications go through the mobile router, performance enhancing solutions are proposed that can be deployed in the mobile router to transparently address the throughput and fairness problems. In this architecture, when the route is known in advance and repetitive (e.g. for public transport or a regularly commuting private vehicle), a certain degree of prediction of impending link disruptions is possible. An anticipatory state freezing mechanism is proposed that relies on the prediction of link disruptions to freeze and unfreeze the state machine of TCP, the widely used transport protocol in the Internet. Simulation study shows that TCP throughput has a non-linear relationship with the prediction accuracy. As prediction accuracy increases, throughput problem diminishes quickly. An adaptive mobile router based fairness control mechanism is proposed to address the unfair sharing of wireless bandwidth in highly dynamic scenarios. The fairness is controlled by dynamically estimating the round-trip-times of all onboard TCP connections and transparently adjusting the protocol control parameters at the router. The thesis also discusses implementation issues for the proposed solutions.

Page generated in 0.2247 seconds