• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 150
  • 20
  • 9
  • 9
  • 3
  • 1
  • 1
  • Tagged with
  • 204
  • 204
  • 194
  • 102
  • 61
  • 34
  • 24
  • 23
  • 23
  • 22
  • 21
  • 20
  • 19
  • 18
  • 17
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
131

Conceptual design of a network architecture for a typical manufacturing information system using open systems integration /

Winig, Robert J. January 1993 (has links)
Thesis (M.S.)--Virginia Polytechnic Institute and State University, 1993. / Vita. Abstract. Includes bibliographical references. Also available via the Internet.
132

Turbo-blast : a novel technique for multi-transmit and multi-receive wireless communications /

Sellathurai, Mathini. January 2001 (has links)
Thesis (Ph.D.) -- McMaster University, 2001. / Includes bibliographical references (leaves 174-186). Also available via World Wide Web.
133

Scheduling for fast turnaround in peer-based desktop grid systems /

Zhou, Dayi, January 2006 (has links)
Thesis (Ph. D.)--University of Oregon, 2006. / Typescript. Includes vita and abstract. Includes bibliographical references (leaves 136-144). Also available for download via the World Wide Web; free to University of Oregon users.
134

An adaptive approach for optimized opportunistic routing over Delay Tolerant Mobile Ad hoc Networks

Zhao, Xiaogeng January 2008 (has links)
This thesis presents a framework for investigating opportunistic routing in Delay Tolerant Mobile Ad hoc Networks (DTMANETs), and introduces the concept of an Opportunistic Confidence Index (OCI). The OCI enables multiple opportunistic routing protocols to be applied as an adaptive group to improve DTMANET routing reliability, performance, and efficiency. The DTMANET is a recently acknowledged networkarchitecture, which is designed to address the challenging and marginal environments created by adaptive, mobile, and unreliable network node presence. Because of its ad hoc and autonomic nature, routing in a DTMANET is a very challenging problem. The design of routing protocols in such environments, which ensure a high percentage delivery rate (reliability), achieve a reasonable delivery time (performance), and at the same time maintain an acceptable communication overhead (efficiency), is of fundamental consequence to the usefulness of DTMANETs. In recent years, a number of investigations into DTMANET routing have been conducted, resulting in the emergence of a class of routing known as opportunistic routing protocols. Current research into opportunistic routing has exposed opportunities for positive impacts on DTMANET routing. To date, most investigations have concentrated upon one or other of the quality metrics of reliability, performance, or efficiency, while some approaches have pursued a balance of these metrics through assumptions of a high level of global knowledge and/or uniform mobile device behaviours. No prior research that we are aware of has studied the connection between multiple opportunistic elements and their influences upon one another, and none has demonstrated the possibility of modelling and using multiple different opportunistic elements as an adaptive group to aid the routing process in a DTMANET. This thesis investigates OCI opportunities and their viability through the design of an extensible simulation environment, which makes use of methods and techniques such as abstract modelling, opportunistic element simplification and isolation, random attribute generation and assignment, localized knowledge sharing, automated scenario generation, intelligent weight assignment and/or opportunistic element permutation. These methods and techniques are incorporated at both data acquisition and analysis phases. Our results show a significant improvement in all three metric categories. In one of the most applicable scenarios tested, OCI yielded a 31.05% message delivery increase (reliability improvement), 22.18% message delivery time reduction (performance improvement), and 73.64% routing depth decrement (efficiency improvement). We are able to conclude that the OCI approach is feasible across a range of scenarios, and that the use of multiple opportunistic elements to aid decision-making processes in DTMANET environments has value.
135

The measurement of enterprise architecture to add value to small and medium enterprises

Ruyter, Masood January 2012 (has links)
Thesis (MTech (Information Technology))--Cape Peninsula University of Technology, 2012. / South Africa has a complex financial and retail service industry with high reliance on the use of IT systems to ensure effectiveness and maintainability. Decision making and improved outcomes may be done through an IT aligned enterprise architecture (EA) strategy. EA is a capability that contributes to the support and success of an organisations' IT. Organisations are currently using EA to better align IT and the business strategy which provides a comprehensive v.ew of the IT system. Thus, EA is increasing in organisations yet the measurement and value of EA is limited to organisations and enterprise architects. The discussions of the benefits and value of EA has been discussed for several years, however there are still no consensus about how the benefits and value of EA can be measured. The lack and clear understanding of the benefits and value of EA needs to consider different aspects of IT as well as the shareholders when measuring the benefits and value of EA to an organisation.
136

Simulação compilada para arquiteturas descritivas em ArchC / Compiled simulation for computer architectures described with ArchC

Bartholomeu, Marcus 11 November 2005 (has links)
Orientadores: Rodolfo Jardim de Azevedo, Guido C. S. Araujo / Tese (doutorado) - Universidade Estadual de Campinas, Instituto de Computação / Made available in DSpace on 2018-08-06T01:58:05Z (GMT). No. of bitstreams: 1 Bartholomeu_Marcus_D.pdf: 1185861 bytes, checksum: 17a0dd5731430c30ae20f0996b1215c1 (MD5) Previous issue date: 2005 / Resumo: o simulador é uma das ferramentas mais importantes para o desenvolvimento de uma nova arquitetura computacional. Entre as vantagens que ele apresenta, destacam-se a flexibilidade e baixo custo. Os primeiros simuladores eram criados manualmente, uma prática muito propensa a erros. Atualmente, Linguagens de Descrição de Arquiteturas (ADLs) facilitam a geração dessas ferramentas. O foco deste trabalho é a pesquisa em técnicas de simulação rápida utilizando a ADL ArchC. Partindo do estado da arte nesta área, a simulação compilada, conseguiu-se melhorar ainda mais o desempenho dos simuladores de conjunto de instruções. Duas otimizações foram propostas. A primeira simula atomicamente os blocos básicos, sem nenhum tipo de teste entre as instruções, oferecendo um ganho de desempenho médio de 70%. A segunda otimização calcula antecipadamente o alvo da maioria das instruções de salto, permitindo o controle do fluxo de execução pelo simulador. Esta otimização é um passo adicional à primeira, oferecendo um ganho de desempenho combinado médio de 180%. Além da simulação rápida, também foi desenvolvida uma metodologia para que simuladores de arquiteturas possam interagir com a máquina hospedeira permitindo, por exemplo, o acesso a arquivos locais e a passagem de parâmetros de linha de commando. Também foi definida uma interface para acesso a dados externos que permitiu a implementação de um protótipo para simulação de arquiteturas com múltiplos processadores. Esta interface de dados será integrada à ArchC e permitirá a co-simulação de blocos genéricos descritos em SystemC / Abstract: The simuIator is one of the most important tools to design a new computer architecture. It has many advantages, the most important are flexibility and Iow cost. The first simulators were created manually, which was an error-prone practice. Nowadays, Architecture Description Languages (ADLs) simplifies the generation of these tools. This work focus on the research of new fast simulation techniques using the ArchC ADL. Beginning from the state-of-art in this area, the compiled simulation, is was possible to speed-up the instruction set simulation performance even higher. Two optimizations were proposed. The first one simulates basic blocks atomically, without any condition test between instructions, and speed-up the simuIation by 70% in average. The second optimization anticipates the majority of target address calculation for jump instructions, allowing the flow control to be done by the simuIator. This second optimization is an improvement to the first one, and provides an speed-up of 180% in average. Besides the research of fast simuIation techniques, a methodology was created to allow architecture simulators to interact with the host machine, which makes it possible, for example, to access local files and take options from command-line. Also, an interface were defined to access external data which allows a propotype imIementation of a multiprocessar architecture simulator. This interface will be integrated to ArchC to achieve co-simulation capability for generic blocks described in SystemC / Doutorado / Doutor em Ciência da Computação
137

Um metodo de testes de integração para sistemas baseados em componentes / A method of integration testing for system based on components

Cardoso, Josiane Aparecida 07 August 2018 (has links)
Orientador: Eliane Martins / Dissertação (mestrado profissional) - Universidade Estadual de Campinas, Instituto de Computação / Made available in DSpace on 2018-08-07T10:05:17Z (GMT). No. of bitstreams: 1 Cardoso_JosianeAparecida_M.pdf: 1727201 bytes, checksum: 2b03579fc8602d15cc944cb59e39086e (MD5) Previous issue date: 2006 / Resumo: O desenvolvimento baseado em componentes vem sendo cada vez mais utilizado hoje em dia, pois permite que um sistema seja construído através da reutilização de software. Um problema associado à reutilização de componentes é o fato de que um componente implementado para um determinado contexto pode ser reaproveitado em um outro contexto com especificações diferentes. Esse tipo de problema pode comprometer a construção de um novo sistema de software confiável baseado em componentes. Falhas devem ser encontradas não só nos componentes, mas na integração dos componentes. No contexto de testes de integração, estratégias têm sido propostas considerando apenas abordagens estáticas com base no diagrama de classes do sistema para se definir uma ordenação para os testes de integração minimizando o número de stubs necessários. Esse trabalho propõe um método de testes de integração para sistemas baseados em componentes que se baseia na arquitetura do sistema considerando uma abordagem dinâmica sem a necessidade do código fonte do componente / Abstract: Nowadays the development based on components is being more frequently used since it allows a system to be constructed through the reuse of software. A problem associated to the reuse of components is that one component implemented for a specific context may be reused in another context with different specifications. This kind of problem may jeopardize the construction of a new reliable software system based on components. Faults must be found not only in the components, but in the integration of the components as well. In the context of integration testing, strategies have been proposed considering only the structural approaches based on the class diagram of the system to define an ordering for the integration testing, minimizing the number of stubs needed. This work proposes a method of integration testing for system based on components, which bases itself on the architecture of the system, considering a behavioral approach without the need for the component source code / Mestrado / Engenharia de Computação / Mestre em Computação
138

Uma arquitetura para negociação de redes virtualizadas na internet do futuro baseada em classes de QoS / A Framework for negotiation of virtualized networks in the future internet based on QoS classes

Gomes, Rafael Lopes, 1987- 07 October 2012 (has links)
Orientador: Edmundo Roberto Mauro Madeira / Dissertação (mestrado) - Universidade Estadual de Campinas, Instituto de Computação / Made available in DSpace on 2018-08-21T07:23:57Z (GMT). No. of bitstreams: 1 Gomes_RafaelLopes_M.pdf: 1407268 bytes, checksum: c68d2355ef384cdcbcdb0658ddb32852 (MD5) Previous issue date: 2012 / Resumo: Ao longo dos anos a Internet vem se tornando o principal meio de comunicação, onde muitas empresas e organizações a usam como base para os seus serviços, sendo que na maioria dos casos, estas empresas têm vários provedores de Internet (Modelo Multi- Provedor). Entretanto, a Internet atual não provê garantias de Qualidade de Serviço (Quality of Service - QoS). Para contornar esse problema, as empresas realizam um Acordo de Nível de Serviços (Service Level Agreement - SLA). Dentro desse contexto, esta dissertação de mestrado tem por objetivo desenvolver uma arquitetura para negociação de redes virtualizadas na Internet do Futuro utilizando técnicas de classificação de tráfegos para decidir, a partir dos dados coletados e das políticas de SLA, por qual ISP (Internet Service Provider) enviar os dados de acordo com a classe que os dados se enquadram. Espera-se assim atender aos requisitos de QoS de cada classe de tráfego, e diminuir os custos da empresa. Para realizar a validação da arquitetura proposta, foram efetuados experimentos baseados no protocolo OpenFlow e no emulador Mininet. Os resultados mostraram a eficiência do modelo desenvolvido, bem como a capacidade do mesmo de cumprir os objetivos definidos / Abstract: Over the years the Internet has become the primary means of communication, where many companies and organizations use it as basis for their services, and in most cases, these companies have multiple Internet service providers (Multi-Provider Model). However, the current Internet does not guarantee Quality of Service (QoS), to circumvent this problem; the companies apply a Service Level Agreements (SLA). Within this context, this dissertation aims to develop architecture for traffic engineering based on traffic classification to decide, from the data collected and SLA policies, for which ISPs to send the data, according to the data class. It is expected to ensure QoS requirements of each traffic class, and reduce the costs of the company. To validate the proposed framework, experiments based on the OpenFlow protocol and on the Mininet emulator were performed. The results showed the efficiency of the framework, as well as its capacity to fulfill the desired requirements / Mestrado / Ciência da Computação / Mestre em Ciência da Computação
139

TOWARDS A SECURITY REFERENCE ARCHITECTURE FOR NETWORK FUNCTION VIRTUALIZATION

Unknown Date (has links)
Network Function Virtualization (NFV) is an emerging technology that transforms legacy hardware-based network infrastructure into software-based virtualized networks. Instead of using dedicated hardware and network equipment, NFV relies on cloud and virtualization technologies to deliver network services to its users. These virtualized network services are considered better solutions than hardware-based network functions because their resources can be dynamically increased upon the consumer’s request. While their usefulness can’t be denied, they also have some security implications. In complex systems like NFV, the threats can come from a variety of domains due to it containing both the hardware and the virtualize entities in its infrastructure. Also, since it relies on software, the network service in NFV can be manipulated by external entities like third-party providers or consumers. This leads the NFV to have a larger attack surface than the traditional network infrastructure. In addition to its own threats, NFV also inherits security threats from its underlying cloud infrastructure. Therefore, to design a secure NFV system and utilize its full potential, we must have a good understanding of its underlying architecture and its possible security threats. Up until now, only imprecise models of this architecture existed. We try to improve this situation by using architectural modeling to describe and analyze the threats to NFV. Architectural modeling using Patterns and Reference Architectures (RAs) applies abstraction, which helps to reduce the complexity of NFV systems by defining their components at their highest level. The literature lacks attempts to implement this approach to analyze NFV threats. We started by enumerating the possible threats that may jeopardize the NFV system. Then, we performed an analysis of the threats to identify the possible misuses that could be performed from them. These threats are realized in the form of misuse patterns that show how an attack is performed from the point of view of attackers. Some of the most important threats are privilege escalation, virtual machine escape, and distributed denial-of-service. We used a reference architecture of NFV to determine where to add security mechanisms in order to mitigate the identified threats. This produces our ultimate goal, which is building a security reference architecture for NFV. / Includes bibliography. / Dissertation (Ph.D.)--Florida Atlantic University, 2020. / FAU Electronic Theses and Dissertations Collection
140

MACHINE LEARNING DEMODULATOR ARCHITECTURES FOR POWER-LIMITED COMMUNICATIONS

Unknown Date (has links)
The success of deep learning has renewed interest in applying neural networks and other machine learning techniques to most fields of data and signal processing, including communications. Advances in architecture and training lead us to consider new modem architectures that allow flexibility in design, continued learning in the field, and improved waveform coding. This dissertation examines neural network architectures and training methods suitable for demodulation in power-limited communication systems, such as those found in wireless sensor networks. Such networks will provide greater connection to the world around us and are expected to contain orders of magnitude more devices than cellular networks. A number of standard and proprietary protocols span this space, with modulations such as frequency-shift-keying (FSK), Gaussian FSK (GFSK), minimum shift keying (MSK), on-off-keying (OOK), and M-ary orthogonal modulation (M-orth). These modulations enable low-cost radio hardware with efficient nonlinear amplification in the transmitter and noncoherent demodulation in the receiver. / Includes bibliography. / Dissertation (Ph.D.)--Florida Atlantic University, 2020. / FAU Electronic Theses and Dissertations Collection

Page generated in 0.0825 seconds