• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 8769
  • 2932
  • 1104
  • 1047
  • 1021
  • 682
  • 315
  • 303
  • 277
  • 266
  • 135
  • 128
  • 79
  • 78
  • 75
  • Tagged with
  • 20131
  • 3915
  • 2833
  • 2577
  • 2439
  • 2350
  • 1937
  • 1840
  • 1554
  • 1536
  • 1518
  • 1512
  • 1502
  • 1445
  • 1399
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
891

Performance of IEEE 802.15.4 beaconless-enabled protocol for low data rate ad hoc wireless sensor networks

Iqbal, Muhamad Syamsu January 2016 (has links)
This thesis focuses on the enhancement of the IEEE 802.15.4 beaconless-enabled MAC protocol as a solution to overcome the network bottleneck, less flexible nodes, and more energy waste at the centralised wireless sensor networks (WSN). These problems are triggered by mechanism of choosing a centralised WSN coordinator to start communication and manage the resources. Unlike IEEE 802.11 standard, the IEEE 802.15.4 MAC protocol does not include method to overcome hidden nodes problem. Moreover, understanding the behaviour and performance of a large-scale WSN is a very challenging task. A comparative study is conducted to investigate the performance of the proposed ad hoc WSN both over the low data rate IEEE 802.15.4 and the high data rate IEEE 802.11 standards. Simulation results show that, in small-scale networks, ad hoc WSN over 802.15.4 outperforms the WSN where it improves 4-key performance indicators such as throughput, PDR, packet loss, and energy consumption by up to 22.4%, 17.1%, 34.1%, and 43.2%, respectively. Nevertheless, WSN achieves less end-to-end delay; in this study, it introduces by up to 2.0 ms less delay than that of ad hoc WSN. Furthermore, the ad hoc wireless sensor networks work well both over IEEE 802.15.4 and IEEE 802.11 protocols in small-scale networks with low traffic loads. The performance of IEEE 802.15.4 declines for the higher payload size since this standard is dedicated to low rate wireless personal access networks. A deep performance investigation of the IEEE 802.15.4 beaconless-enabled wireless sensor network (BeWSN) in hidden nodes environment has been conducted and followed by an investigation of network overhead on ad hoc networks over IEEE 802.11 protocol. The result of investigation evinces that the performance of beaconless-enabled ad hoc wireless sensor networks deteriorates as indicated by the degradation of throughput and packet reception by up to 72.66 kbps and 35.2%, respectively. In relation to end-to-end delay, however, there is no significant performance deviation caused by hidden nodes appearance. On the other hand, preventing hidden node effect by implementing RTS/CTS mechanism introduces significant overhead on the network that applies low packet size. Therefore, this handshaking method is not suitable for low rate communications protocol such as IEEE 802.15.4 standard. An evaluation study of a 101-node large-scale beaconless-enabled wireless sensor networks over IEEE 802.15.4 protocol has been carried out after the nodes deployment model was arranged. From the experiment, when the number of connection densely increases, then the probability of packet delivery decreases by up to 40.5% for the low payload size and not less than of 44.5% for the upper payload size. Meanwhile, for all sizes of payload applied to the large-scale ad hoc wireless sensor network, it points out an increasing throughput whilst the network handles more connections among sensor nodes. In term of dropped packet, it confirms that a fewer data drops at smaller number of connecting nodes on the network where the protocol outperforms not less than of 34% for low payload size of 30 Bytes. The similar trend obviously happens on packet loss. In addition, the simulation results show that the smaller payload size performs better than the bigger one in term of network latency, where the payload size of 30 Bytes contributes by up to 41.7% less delay compared with the contribution of the payload size of 90 Bytes.
892

Analyzing network monitoring systems and objects for a telecommunications company

Arvedal, David January 2017 (has links)
The goal with this thesis work has been to identify what a telecommunications company should monitor and to find a network monitoring system that can monitor these identified objects on two different platforms: Windows and Linux. The network monitoring system has been implemented in a telecommunications company’s environment and this thesis presents how the system monitors their environment. The subject for this thesis work is within network monitoring. The problem formulation has been answered by conducting a literature study and by testing network monitoring systems’ features in a lab environment. The sources used in the literature study consists of scientific articles and other articles found on the web. The lab environment consisted of virtual machines that runs Linux or Windows as an operating system.   The purpose of the work was to enlighten Cellip in what objects they should monitor and to help the company to monitor them by implementing a network monitoring system. Cellip is a telecommunications company that provides IP-telephony services through Session Initiation Protocol. The limits of this thesis work are based on what their environment supports in terms of monitoring. Cellip’s environment consists of Linux and Windows servers, Cisco switches and firewalls, and Sonus Session Border Controllers.   In summary, the result of this thesis gives the reader information about what a telecommunications company with a similar environment to Cellip should monitor, what three systems that can monitor these objects, which of the three systems that has most automatized features and finally how the chosen system Datadog monitors and presents the objects. Some of the objects that are important to monitor is: memory, disk storage, latency, packet loss. In conclusion, this thesis presents a monitoring baseline for telecommunication companies with a similar environment to Cellip.
893

Uma abordagem baseada em aspectos topológicos para expansão de redes físicas no contexto de virtualização de redes / An approach based on topological factors for the expansion of physical infrastructure in the context of network virtualization

Luizelli, Marcelo Caggiani January 2014 (has links)
A virtualização de redes é um mecanismo que permite a coexistência de múltiplas redes virtuais sobre um mesmo substrato físico. Um dos desafios de pesquisa abordados na literatura é o mapeamento eficiente de recursos virtuais em infraestruturas físicas. Embora o referido desafio tenha recebido considerável atenção, as abordagens que constituem o estado-da-arte apresentam alta taxa de rejeição, i.e., a proporção de solicitações de redes virtuais negadas em relação ao total de solicitações efetuadas ao substrato é elevada. Nesta dissertação, caracteriza-se, inicialmente, a relação entre a qualidade dos mapeamentos de redes virtuais e as estruturas topológicas dos substratos subjacentes. Avalia-se as soluções exatas de um modelo de mapeamento online sob diferentes classes de topologias de rede. A partir do entendimento dos fatores topológicos que influenciam diretamente o processo de mapeamento de redes virtuais, propõe-se uma estratégia para planejar a expansão de redes de provedores de infraestrutura de forma a reduzir consistentemente a taxa de rejeição de requisições de redes virtuais e melhor aproveitar os recursos ociosos da mesma. Os resultados obtidos evidenciam que grande parte das rejeições de redes virtuais ocorre em situações em que há grande disponibilidade de recursos, mas alguns poucos já saturados acabam inviabilizando, em função de características de conectividade do substrato, o atendimento de novas requisições. Ademais, os resultados obtidos utilizando a estratégia proposta evidenciam que o fortalecimento de partes-chave da infraestrutura levam a uma ocupação muito mais satisfatória. Uma expansão de 10% a 20% dos recursos da infraestrutura contribui para um aumento sustentado de até 30% no número de redes virtuais aceitas e de até 45% no aproveitamento dos recursos em comparação com a rede original. / Network virtualization is a mechanism that allows the coexistence of multiple virtual networks on top of a single physical substrate. One of the research challenges addressed recently in the literature is the efficient mapping of virtual resources on physical infrastructures. Although this challenge has received considerable attention, state-of-the-art approaches present, in general, a high rejection rate, i.e., the ratio between the number of denied virtual network requests and the total amount of requests is considerably high. In this thesis, we characterize the relationship between the quality of virtual network mappings and the topological structures of the underlying substrates. Exact solutions of an online embedding model are evaluated under different classes of network topologies. From the understanding of the topological factors that directly influence the virtual network embedding process, we propose an expansion strategy of physical infrastructure in order to suggest adjustments that lead to higher virtual network acceptance and, in consequence, to improved physical resource utilization. The obtained results demonstrate that most of rejections occur in situations in which a significant amount of resource is available, but a few saturated devices and links, depending on connectivity features of the physical substrate, hinder the acceptance of new requests. Moreover, the obtained results using the proposed strategy evidence that an expansion of 10% to 20% of the infrastructure resources leads to a sustained increase of up to 30% in the number of accepted virtual networks and of up to 45% in resource usage compared to the original network.
894

Foundational Forensic Techniques for Cellular and Ad Hoc Multi-hop Networks

Zhao, Xiwei 26 March 2008 (has links)
The Internet has become an integral part of our nation's critical socio-economic infrastructure. With its heightened use and growing complexity however, organizations are at greater risk of cyber crimes. To aid in the investigation of crimes committed on or via the Internet, a network forensics analysis tool pulls together needed digital evidence. It provides a platform for performing deep network analysis by capturing, recording and analyzing network events to find out the source of a security attack or other information security incidents. Existing network forensics work has been mostly focused on the Internet and fixed networks. But the exponential growth and use of wireless technologies, coupled with their unprecedented characteristics, necessitates the development of new network forensic analysis tools. This dissertation fostered the emergence of a new research field in cellular and ad-hoc network forensics. It was one of the first works to identify this problem and offer fundamental techniques and tools that laid the groundwork for future research. In particular, it introduced novel methods to record network incidents and report logged incidents. For recording incidents, location is considered essential to documenting network incidents. However, in network topology spaces, location cannot be measured due to absence of a 'distance metric'. Therefore, a novel solution was proposed to label locations of nodes within network topology spaces, and then to authenticate the identity of nodes in ad hoc environments. For reporting logged incidents, a novel technique based on Distributed Hash Tables (DHT) was adopted. Although the direct use of DHTs for reporting logged incidents would result in an uncontrollably recursive traffic, a new mechanism was introduced that overcome this recursive process. These logging and reporting techniques aided forensics over cellular and ad-hoc networks, which in turn increased their ability to track and trace attacks to their source. These techniques were a starting point for further research and development that would result in equipping future ad hoc networks with forensic components to complement existing security mechanisms.
895

Wide Activated Separate 3D Convolution for Video Super-Resolution

Yu, Xiafei 18 December 2019 (has links)
Video super-resolution (VSR) aims to recover a realistic high-resolution (HR) frame from its corresponding center low-resolution (LR) frame and several neighbouring supporting frames. The neighbouring supporting LR frames can provide extra information to help recover the HR frame. However, these frames are not aligned with the center frame due to the motion of objects. Recently, many video super-resolution methods based on deep learning have been proposed with the rapid development of neural networks. Most of these methods utilize motion estimation and compensation models as preprocessing to handle spatio-temporal alignment problem. Therefore, the accuracy of these motion estimation models are critical for predicting the high-resolution frames. Inaccurate results of motion compensation models will lead to artifacts and blurs, which also will damage the recovery of high-resolution frames. We propose an effective wide activated separate 3 dimensional (3D) Convolution Neural Network (CNN) for video super-resolution to overcome the drawback of utilizing motion compensation models. Separate 3D convolution factorizes the 3D convolution into convolutions in the spatial and temporal domain, which have benefit for the optimization of spatial and temporal convolution components. Therefore, our method can capture temporal and spatial information of input frames simultaneously without additional motion evaluation and compensation model. Moreover, the experimental results demonstrated the effectiveness of the proposed wide activated separate 3D CNN.
896

An extended Bayesian network approach for analyzing supply chain disruptions

Donaldson Soberanis, Ivy Elizabeth 01 January 2010 (has links)
Supply chain management (SCM) is the oversight of materials, information, and finances as they move in a process from supplier to manufacturer to wholesaler to retailer to consumer. Supply chain management involves coordinating and integrating these flows both within and among companies as efficiently as possible. The supply chain consists of interconnected components that can be complex and dynamic in nature. Therefore, an interruption in one subnetwork of the system may have an adverse effect on another subnetworks, which will result in a supply chain disruption. Disruptions from an event or series of events can have costly and widespread ramifications. When a disruption occurs, the speed at which the problem is discovered becomes critical. There is an urgent need for efficient monitoring of the supply chain. By examining the vulnerability of the supply chain network, supply chain managers will be able to mitigate risk and develop quick response strategies in order to reduce supply chain disruption. However, modeling these complex supply chain systems is a challenging research task. This research is concerned with developing an extended Bayesian Network approach to analyze supply chain disruptions. The aim is to develop strategies that can reduce the adverse effects of the disruptions and hence improve overall system reliability. The supply chain disruptions is modeled using Bayesian Networks-a method of modeling the cause of current and future events, which has the ability to model the large number of variables in a supply chain and has proven to be a powerful tool under conditions of uncertainty. Two impact factors are defined. These are the Bayesian Impact Factor (BIF) and the Node Failure Impact Factor (NFIF). An industrial example is used to illustrate the application proposed to make the supply chain more reliable.
897

Návrh realizace počítačové sítě pro střední školu / High School Computer Network Realization Design

Kožiak, Ján January 2018 (has links)
The aim of this thesis is to create a computer network for Secondary Professional School in Tisovec. In the theoretical part we explain the terms connected to our issue. The practical part focuses on the creation of a computer network. This part includes a selection of suitable network components, graphical display of network wiring and data connector arrangement. The thesis deal with the project creation, thus in the practical part we describe computer network from a project management point of view.
898

Plánování přístupových sítí UMTS / Planning of UTMS access networks

Spěváček, Jiří January 2008 (has links)
The master thesis is focused on the Universal Mobile Telecommunications System (UMTS) - the third-generation cell phone technology. The first part describes the development of mobile networks and the structure of the UMTS. The second part, the main part of the thesis, is focused on the planning procedure of the UMTS access network. It consists of an explanation of the general planning procedure and a specific project designing the coverage of Pelhřimov surroundings. The final part of the thesis contains a laboratory problem including instructions.
899

Možnosti využití neuronových sítí v síťových prvcích / Potential application of neural networks in network elements

Babnič, Patrik January 2011 (has links)
The goal was to get acquainted with the problems of network elements to describe neural networks that can be used to manage such a feature. The theoretical part deals with the neural networks from their inception to the present. It focuses mainly on the network, witch can be used for management control. These are the two network: Hopfield network and Kohonen network. The practical part deals with the network element model and ist implementation. It contains a practical element model using a neural network, witch is controlled by a network element.
900

Attacking and securing Network Time Protocol

Malhotra, Aanchal 14 February 2020 (has links)
Network Time Protocol (NTP) is used to synchronize time between computer systems communicating over unreliable, variable-latency, and untrusted network paths. Time is critical for many applications; in particular it is heavily utilized by cryptographic protocols. Despite its importance, the community still lacks visibility into the robustness of the NTP ecosystem itself, the integrity of the timing information transmitted by NTP, and the impact that any error in NTP might have upon the security of other protocols that rely on timing information. In this thesis, we seek to accomplish the following broad goals: 1. Demonstrate that the current design presents a security risk, by showing that network attackers can exploit NTP and then use it to attack other core Internet protocols that rely on time. 2. Improve NTP to make it more robust, and rigorously analyze the security of the improved protocol. 3. Establish formal and precise security requirements that should be satisfied by a network time-synchronization protocol, and prove that these are sufficient for the security of other protocols that rely on time. We take the following approach to achieve our goals incrementally. 1. We begin by (a) scrutinizing NTP's core protocol (RFC 5905) and (b) statically analyzing code of its reference implementation to identify vulnerabilities in protocol design, ambiguities in specifications, and flaws in reference implementations. We then leverage these observations to show several off- and on-path denial-of-service and time-shifting attacks on NTP clients. We then show cache-flushing and cache-sticking attacks on DNS(SEC) that leverage NTP. We quantify the attack surface using Internet measurements, and suggest simple countermeasures that can improve the security of NTP and DNS(SEC). 2. Next we move beyond identifying attacks and leverage ideas from Universal Composability (UC) security framework to develop a cryptographic model for attacks on NTP's datagram protocol. We use this model to prove the security of a new backwards-compatible protocol that correctly synchronizes time in the face of both off- and on-path network attackers. 3. Next, we propose general security notions for network time-synchronization protocols within the UC framework and formulate ideal functionalities that capture a number of prevalent forms of time measurement within existing systems. We show how they can be realized by real-world protocols (including but not limited to NTP), and how they can be used to assert security of time-reliant applications-specifically, cryptographic certificates with revocation and expiration times. Our security framework allows for a clear and modular treatment of the use of time in security-sensitive systems. Our work makes the core NTP protocol and its implementations more robust and secure, thus improving the security of applications and protocols that rely on time.

Page generated in 0.1005 seconds