• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 777
  • 217
  • 122
  • 65
  • 54
  • 33
  • 32
  • 30
  • 28
  • 21
  • 15
  • 14
  • 9
  • 9
  • 7
  • Tagged with
  • 1595
  • 1595
  • 390
  • 281
  • 244
  • 242
  • 235
  • 231
  • 231
  • 226
  • 215
  • 210
  • 176
  • 173
  • 152
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
481

What Are the Security Challenges Concerning Maintenance Data in the Railway Industry

Khan, Hiba January 2019 (has links)
Recently, technology advancement has brought improvement in all the sectors, including the railway sector. The Internet of Things (IoT) based railway systems have immense potential to improve quality and systems that will enable more efficient, environmental friendly railway system. Many research brought innovations that offer enormous benefits for rail travel. The current research focuses on the railway industries, as they want to reap the benefits of IT concept such as Cloud Computing, Information Security, and Internet of Things (IoT). Railway industries are generating a large volume of data every day from different sources. In addition, machine and human interactions are rapidly increasing along with the development of technologies. This data need to be properly gathered, analysed and shared in a way that it is safe from different types of cyberattacks and calamities. To overcome smart devices’ and Cloud’s limitations, the new paradigm known as Fog computing has appeared. In which an additional layer processes the data and sends the results to the Cloud. Despite numerous benefits of Fog, computing brings into IoT-based environments, privacy and security issues remain the main challenge for its implementation. Hence, the primary purpose of this research is to investigate the potential challenges, consequences, threats, vulnerabilities, and risk management of data security in the railway infrastructure in the context of eMaintenance.
482

Determinants influencing adoption of cloud computing by small medium enterprises in South Africa

Matandela, Wanda January 2017 (has links)
Submitted in partial fulfillment of the requirements for the degree of Master of Commerce in Information Systems (Coursework) at the School of Economic and Business Sciences, University of the Witwatersrand, June 2017 / Small Medium Enterprises (SMEs) are now recognized as the driving force behind most thriving economies. This is mainly attributed to the role they play in most economies in reducing unemployment and their contribution towards Gross Domestic Product. This means that SMEs should have the right resources to enable them to enhance performance. Choosing the right technology to support their businesses is one of the important decisions that SMEs should make. Understanding the benefits and challenges of different technologies is often a problem for most SMEs. One of the new technologies that has gained prominence in recent years is cloud computing. Even though the value associated with this technology has been widely researched especially for large enterprises, the rate at which SMEs adopt cloud computing still remains low. The purpose of this research sought to explore and describe the determinants influencing the adoption of cloud computing by SMEs in South Africa. The study used Technology Organization Environment (TOE) framework as the theoretical lens in understanding the adoption of Could Computing phenomenon. Further, this qualitative exploratory and descriptive study used semi-structured interviews to collect data from five SMEs based in Johannesburg, Gauteng Province, operating in different industries and belonging to the National Small Business Chamber. The main factors that were identified as playing an important role in the adoption of cloud computing by SMEs are, relative advantage, complexity, compatibility, awareness, trialability, culture, top management support, size, regulation and trade partner relationship. It is worth noting that there was not enough evidence that competitive pressure played a significant role in SME cloud adoption. / XL2018
483

A collaborative architecture against DDOS attacks for cloud computing systems. / Uma arquitetura colaborativa contra ataques distribuídos de negação de serviço para sistemas de computação em nuvem.

Almeida, Thiago Rodrigues Meira de 14 December 2018 (has links)
Distributed attacks, such as Distributed Denial of Service (DDoS) ones, require not only the deployment of standalone security mechanisms responsible for monitoring a limited portion of the network, but also distributed mechanisms which are able to jointly detect and mitigate the attack before the complete exhaustion of network resources. This need led to the proposal of several collaborative security mechanisms, covering different phases of the attack mitigation: from its detection to the relief of the system after the attack subsides. It is expected that such mechanisms enable the collaboration among security nodes through the distributed enforcement of security policies, either by installing security rules (e.g., for packet filtering) and/or by provisioning new specialized security nodes on the network. Albeit promising, existing proposals that distribute security tasks among collaborative nodes usually do not consider an optimal allocation of computational resources. As a result, their operation may result in a poor Quality of Service for legitimate packet flows during the mitigation of a DDoS attack. Aiming to tackle this issue, this work proposes a collaborative solution against DDoS attacks with two main goals: (1) ensure an optimal use of resources already available in the attack\'s datapath in a proactive way, and (2) optimize the placement of security tasks among the collaborating security nodes. Regardless the characteristics of each main goal, legitimate traffic must be preserved as packet loss is reduced as much as possible. / Sem resumo
484

Modeling and analysis of security

Unknown Date (has links)
Cloud Computing is a new computing model consists of a large pool of hardware and software resources on remote datacenters that are accessed through the Internet. Cloud Computing faces significant obstacles to its acceptance, such as security, virtualization, and lack of standardization. For Cloud standards, there is a long debate about their role, and more demands for Cloud standards are put on the table. The Cloud standardization landscape is so ambiguous. To model and analyze security standards for Cloud Computing and web services, we have surveyed Cloud standards focusing more on the standards for security, and we classified them by groups of interests. Cloud Computing leverages a number of technologies such as: Web 2.0, virtualization, and Service Oriented Architecture (SOA). SOA uses web services to facilitate the creation of SOA systems by adopting different technologies despite their differences in formats and protocols. Several committees such as W3C and OASIS are developing standards for web services; their standards are rather complex and verbose. We have expressed web services security standards as patterns to make it easy for designers and users to understand their key points. We have written two patterns for two web services standards; WS-Secure Conversation, and WS-Federation. This completed an earlier work we have done on web services standards. We showed relationships between web services security standards and used them to solve major Cloud security issues, such as, authorization and access control, trust, and identity management. Close to web services, we investigated Business Process Execution Language (BPEL), and we addressed security considerations in BPEL and how to enforce them. To see how Cloud vendors look at web services standards, we took Amazon Web Services (AWS) as a case-study. By reviewing AWS documentations, web services security standards are barely mentioned. We highlighted some areas where web services security standards could solve some AWS limitations, and improve AWS security process. Finally, we studied the security guidance of two major Cloud-developing organizations, CSA and NIST. Both missed the quality of attributes offered by web services security standards. We expanded their work and added benefits of adopting web services security standards in securing the Cloud. / Includes bibliography. / Dissertation (Ph.D.)--Florida Atlantic University, 2013.
485

Predictive modeling for chronic conditions

Unknown Date (has links)
Chronic Diseases are the major cause of mortality around the world, accounting for 7 out of 10 deaths each year in the United States. Because of its adverse effect on the quality of life, it has become a major problem globally. Health care costs involved in managing these diseases are also very high. In this thesis, we will focus on two major chronic diseases Asthma and Diabetes, which are among the leading causes of mortality around the globe. It involves design and development of a predictive analytics based decision support system which uses five supervised machine learning algorithm to predict the occurrence of Asthma and Diabetes. This system helps in controlling the disease well in advance by selecting its best indicators and providing necessary feedback. / Includes bibliography. / Thesis (M.S.)--Florida Atlantic University, 2015. / FAU Electronic Theses and Dissertations Collection
486

Projeto de um broker de gerenciamento adaptativo de recursos em computação em nuvem baseado em técnicas de controle realimentado / Design of an adaptive resource management broker for cloud computing based on feedback control techniques

Nobile, Pedro Northon 25 February 2013 (has links)
Computação em nuvem refere-se a um modelo de disponibilização de recursos computacionais no qual a infraestrutura de software e hardware é ofertada como um serviço, e vem se estabelecendo como um paradigma de sucesso graças a versatilidade e ao custo-efetividade envolvidos nesse modelo de negócio, possibilitando o compartilhamento de um conjunto de recursos físicos entre diferentes usuários e aplicações. Com o advento da computação em nuvem e a possibilidade de elasticidade dos recursos computacionais virtualizados, a alocação dinâmica de recursos vem ganhando destaque, e com ela as questões referentes ao estabelecimento de contratos e de de qualidade de serviço. Historicamente, as pesquisas em QoS concentram-se na solução de problemas que envolvem duas entidades: usuários e servidores. Entretanto, em ambientes de nuvem, uma terceira entidade passa a fazer parte dessa interação, o consumidor de serviços em nuvem, que usa a infraestrutura para disponibilizar algum tipo de serviço aos usuários finais e que tem recebido pouca atenção das pesquisa até o momento, principalmente no que tange ao desenvolvimento de mecanismos automáticos para a alocação dinâmica de recursos sob variação de demanda. Este trabalho consiste na proposta de uma arquitetura de gerenciamento adaptativo de recursos sob a perspectiva do modelo de negócio envolvendo três entidades, focada na eficiência do consumidor. O trabalho inspira-se em técnicas de controle realimentado para encontrar soluções adaptativas aos problemas de alocação dinâmica de recursos, resultando em uma arquitetura de broker de consumidor, um respectivo protótipo e um método de projeto de controle para sistemas computacionais dessa natureza / CLoud computing refers to a computer resource deployment model in which software and hardware infrastructure are offered as a service. Cloud computing has become a successful paradigm due to the versatility and cost-effectiveness involved in that business model, making it possible to share a cluster of physical resources between several users and applications. With the advent of cloud computing and the computer elastic resource, dynamic allocation of virtualized resources is becoming more prominent, and along with it, the issues concerning the establishment of quality of service parameters. Historically, research on QoS has focused on solutions for problems involving two entities: users and servers. However, in cloud environments, a third party becomes part of this interaction, the cloud consumer, that uses the infrastructure to provide some kind of service to endusers, and which has received fewer attention, especially regarding the development of autonomic mechanisms for dynamic resource allocation under time-varying demand. This work aims at the development of an architecture for dynamic adaptive resource allocation involving three entities, focused on consumer revenue. The research outcome is a consumer broker architecture based on feedback control, a respective architecture prototype and a computer system feedback control methodology which may be applied in this class of problems
487

Orquestração de migração massiva de máquinas virtuais baseada em análise cíclica para ambientes de computação na nuvem. / Massive virtual machine live migration orchestration for cloud computing environment based on cyclic analyses.

Baruchi, Artur 15 May 2015 (has links)
Um das principais características da tecnologia de virtualização é a Live Migration, que permite que máquinas virtuais sejam movimentadas entre máquinas físicas sem a interrupção da execução. Esta característica habilita a implementação de políticas mais sofisticadas dentro de um ambiente de computação na nuvem, como a otimização de uso de energia elétrica e recursos computacionais. Entretanto, a Live Migration pode impor severa degradação de desempenho nas aplicações das máquinas virtuais e causar diversos impactos na infraestrutura dos provedores de serviço, como congestionamento de rede e máquinas virtuais co-existentes nas máquinas físicas. Diferente de diversos estudos, este estudo considera a carga de trabalho da máquina virtual um importante fator e argumenta que escolhendo o momento adequado para a migração da máquina virtual pode-se reduzir as penalidades impostas pela Live Migration. Este trabalho introduz a Application-aware Live Migration (ALMA), que intercepta as submissões de Live Migration e, baseado na carga de trabalho da aplicação, adia a migração para um momento mais favorável. Os experimentos conduzidos neste trabalho mostraram que a arquitetura reduziu em até 74% o tempo das migrações para os experimentos com benchmarks e em até 67% os experimentos com carga de trabalho real. A transferência de dados causada pela Live Migration foi reduzida em até 62%. Além disso, o presente introduz um modelo que faz a predição do custo da Live Migration para a carga de trabalho e também um algoritmo de migração que não é sensível à utilização de memória da máquina virtual. / A key feature in virtualization technology is the Live Migration, which allows a Virtual Machine to be moved from a physical host to another without execution interruption. This feature enables the implementation of more sophisticated policies inside a cloud environment, such as energy and computational resources optimization. However, live migration can impose severe performance degradation for virtual machine application and cause multiple impacts in service provider infrastructure, such as network congestion and co-located virtual machine performance degradation. Unlike of several studies this work consider the virtual machine workload an important factor and argue that carefully choosing a proper moment to migrate it can reduce the live migration penalties. This work introduces the Application-aware Live Migration Architecture (ALMA), which intercepts live migrations submissions and, based in the application workload, postpone the migration to a more propitious live migration moment. The experiments conducted by this work demonstrated that the architecture reduced up to 74% for live migration time for benchmarks and 67% for real applications workload. The network data transfer during the live migration was reduced up to 62%. Also, the present work introduces a model to predict live migration cost for the application and an algorithm that it is not memory usage sensitive.
488

Informações de suporte ao escalonamento de workflows científicos para a execução em plataformas de computação em nuvem / Support information to scientific workflow scheduling for execution in cloud computing platforms

Teixeira, Eduardo Cotrin 26 April 2016 (has links)
A ciência tem feito uso frequente de recursos computacionais para execução de experimentos e processos científicos, que podem ser modelados como workflows que manipulam grandes volumes de dados e executam ações como seleção, análise e visualização desses dados segundo um procedimento determinado. Workflows científicos têm sido usados por cientistas de várias áreas, como astronomia e bioinformática, e tendem a ser computacionalmente intensivos e fortemente voltados à manipulação de grandes volumes de dados, o que requer o uso de plataformas de execução de alto desempenho como grades ou nuvens de computadores. Para execução dos workflows nesse tipo de plataforma é necessário o mapeamento dos recursos computacionais disponíveis para as atividades do workflow, processo conhecido como escalonamento. Plataformas de computação em nuvem têm se mostrado um alternativa viável para a execução de workflows científicos, mas o escalonamento nesse tipo de plataforma geralmente deve considerar restrições específicas como orçamento limitado ou o tipo de recurso computacional a ser utilizado na execução. Nesse contexto, informações como a duração estimada da execução ou limites de tempo e de custo (chamadas aqui de informações de suporte ao escalonamento) são importantes para garantir que o escalonamento seja eficiente e a execução ocorra de forma a atingir os resultados esperados. Este trabalho identifica as informações de suporte que podem ser adicionadas aos modelos de workflows científicos para amparar o escalonamento e a execução eficiente em plataformas de computação em nuvem. É proposta uma classificação dessas informações, e seu uso nos principais Sistemas Gerenciadores de Workflows Científicos (SGWC) é analisado. Para avaliar o impacto do uso das informações no escalonamento foram realizados experimentos utilizando modelos de workflows científicos com diferentes informações de suporte, escalonados com algoritmos que foram adaptados para considerar as informações inseridas. Nos experimentos realizados, observou-se uma redução no custo financeiro de execução do workflow em nuvem de até 59% e redução no makespan chegando a 8,6% se comparados à execução dos mesmos workflows sendo escalonados sem nenhuma informação de suporte disponível. / Science has been using computing resources to perform scientific processes and experiments that can be modeled as workflows handling large data volumes and performing actions such as selection, analysis and visualization of these data according to a specific procedure. Scientific workflows have been used by scientists from many areas, such as astronomy and bioinformatics, and tend to be computationally intensive and heavily focused on handling large data volumes, which requires using high-performance computing platforms such as grids or clouds. For workflow execution in these platforms it is necessary to assign the workflow activities to the available computational resources, a process known as scheduling. Cloud computing platforms have proved to be a viable alternative for scientific workflows execution, but scheduling in cloud must take into account specific constraints such as limited budget or the type of computing resources to be used in execution. In this context, information such as the estimated duration of execution, or time and cost limits (here this information is generally referred to as scheduling support information) become important for efficient scheduling and execution, aiming to achieve the expected results. This work identifies support information that can be added to scientific workflow models to support efficient scheduling and execution in cloud computing platforms. We propose and analyze a classification of such information and its use in Scientific Workflows Management Systems (SWMS). To assess the impact of support information on scheduling, experiments were conducted with scientific workflow models using different support information, scheduled with algorithms that were adapted to consider the added information. The experiments have shown a reduction of up to 59% on the financial cost of workflow execution in the cloud, and a reduction reaching 8,6% on the makespan if compared to workflow execution scheduled without any available supporting information.
489

Metodologia de benchmark para avaliação de desempenho não-estacionária: um estudo de caso baseado em aplicações de computação em nuvem / Benchmark methodology for non-stationary performance evaluation: a case study based on cloud computing applications

Mamani, Edwin Luis Choquehuanca 23 February 2016 (has links)
Este trabalho analisa os efeitos das propriedades dinâmicas de sistemas distribuídos de larga escala e seu impacto no desempenho, e introduz uma abordagem para o planejamento de experimentos de referência capazes de expor essa influência. Especialmente em aplicações complexas com múltiplas camadas de software (multi-tier), o efeito total de pequenos atrasos introduzidos por buffers, latência na comunicação e alocação de recursos, pode resultar gerando inércia significativa ao longo do funcionamento do sistema. A fim de detetar estas propriedade dinâmica, o experimento de execução do benchmark deve excitar o sistema com carga de trabalho não-estacionária sob um ambiente controlado. A presente pesquisa discorre sobre os elementos essenciais para este fim e ilustra a abordagem de desenvolvimento com um estudo de caso. O trabalho também descreve como a metodologia de instrumentação pode ser explorada em abordagens de modelagem dinâmica para sobrecargas transientes devido a distúrbios na carga de trabalho. / This work examines the effects of dynamic properties of large-scale distributed systems and their impact on the delivered performance, and introduces an approach to the design of benchmark experiments capable of exposing this influence. Specially in complex, multi-tier applications, the net effect of small delays introduced by buffers, communication latency and resource instantiation, may result in significant inertia along the input-output path. In order to bring out these dynamic property, the benchmark experiment should excite the system with non-stationary workload under controlled conditions. The present research report elaborates on the essentials for this purpose and illustrates the design approach through a case study. The work also outlines how the instrumentation methodology can be exploited in dynamic modeling approaches to model transient overloads due to workload disturbances.
490

Security architecture for network virtualization in cloud computing. / Arquitetura de segurança para virtualização de redes em computação em nuvem.

Barros, Bruno Medeiros de 10 October 2016 (has links)
Network virtualization has been a quite active research area in the last years, aiming to tackle the increasing demand for high performance and secure communication in cloud infrastructures. In special, such research eforts have led to security solutions focused on improving isolation among multiple tenant of public clouds, an issue recognized as critical both by the academic community and by the technology Industry. More recently, the advent of Software-Defined Networks (SDN) and of Network Function Virtualization (NFV) introduced new concepts and techniques for addressing issues related to the isolation of network resources in multi-tenant clouds while improving network manageability and flexibility. Similarly, hardware technologies such as Single Root I/O Virtualization (SR-IOV) enable network isolation in the hardware level while improving performance in physical and virtual networks. Aiming to provide a cloud network environment that effciently tackles multi-tenant isolation, we present three complementary strategies for addressing the isolation of resources in cloud networks. These strategies are then applied in the evaluation of existing network virtualization architectures, exposing the security gaps associated to current technologies, and paving the path for novel solutions. We then propose a security architecture that builds upon the strategies presented, as well as upon SDN, NFV and SR-IOV technologies, to implement secure cloud network domains. The theoretical and experimental analyses of the resulting architecture show a considerable reduction of the attack surface in tenant networks, with a small impact over tenants\' intra-domain and inter-domain communication performance. / Virtualização de redes é uma área de pesquisa que tem ganho bastante atenção nos últimos anos, motivada pela necessidade de se implementar sistemas de comunicação seguros e de alta performance em infraestruturas de computação em nuvem. Em particular, os esforços de pesquisa nesta área têm levado ao desenvolvimento de soluções de segurança que visam aprimorar o isolamento entre os múltiplos inquilinos de sistemas de computação em nuvem públicos, uma demanda reconhecidamente crítica tanto pela comunidade acadêmica quanto pela indústria de tecnologia. Mais recentemente, o advento das Redes Definidas por Software (do inglês Software-Defined Networks - SDN) e da Virtualização de Funções de Rede (do inglês Network Function Virtualization - NFV) introduziu novos conceitos e técnicas que podem ser utilizadas para abordar questões de isolamento de redes virtualizadas em sistemas de computação em nuvem com múltiplos inquilinos, enquanto aprimoram a capacidade de gerenciamento e a flexibilidade de suas redes. Similarmente, tecnologias de virtualização assistida por hardware como Single Root I/O Virtualization - SR-IOV permitem a implementação do isolamento de recursos de hardware, melhorando o desempenho de redes físicas e virtualizadas. Com o intuito de implementar uma solução de virtualização de redes que aborda de maneira eficiente o problema de isolamento entre múltiplos inquilinos, nós apresentamos três estratégias complementares para o isolamento de recursos de rede em sistemas computação em nuvem. As estratégias apresentadas são então aplicadas na avaliação de arquiteturas de virtualização de rede existentes, revelando lacunas de segurança associadas às tecnologias utilizadas atualmente, e abrindo caminho para o desenvolvimento de novas soluções. Nós então propomos uma arquitetura de segurança que utiliza as estratégias apresentadas, e tecnologias como SDN, NFV e SR-IOV, para implementar domínios de rede seguros. As análises teórica e experimental da arquitetura proposta mostram considerável redução das superfícies de ataque em redes virtualizadas, com um pequeno impacto sobre o desempenho da comunicação entre máquinas virtuais de inquilinos da nuvem.

Page generated in 0.0792 seconds