• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 4
  • 4
  • 1
  • Tagged with
  • 10
  • 10
  • 4
  • 4
  • 4
  • 3
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

Aspects of optimal sequential resource allocation

Ferreira Pena Do Amaral, J. A. January 1985 (has links)
No description available.
2

Resource allocation for cooperative cognitive radios

Lessinnes, Mathieu 20 January 2014 (has links)
Resource allocation consists in allocating spectrum and power on every link of a network, possibly under power and rate requirements. In the context of cognitive radios, almost 15 years of research produced an impressive amount of theoretical contributions, exploring a wide range of possibilities. However, despite the ever-growing list of imaginable scenarios, we observe in Chapter 2 that most of these studies are based on similar working hypotheses. Our first contribution is to challenge some of these hypotheses, and propose a novel resource allocation scheme. Sticking to realistic assumptions, we show how our scheme reduces both computational complexity and control traffic, compared to other state-of-the-art techniques.<p><p>Due to a majority of the abovementioned studies making some constraining assumptions, realistic system designs and experimental demonstrations are much more quiet and unharvested fields. In an effort to help this transition from theory to practice, our second contribution is a four-nodes cognitive network demonstrator, presented in Chapter 3. In particular, we aim at providing a modular platform available for further open collaboration: different options for spectrum sensing, resource allocation, synchronisation and others can be experimented on this demonstrator. As an example, we develop a simple protocol to show that our proposed resource allocation scheme is fully implementable, and that primary users can be avoided using our approach.<p><p>Chapter 4 aims at removing another working hypothesis made when developping our resource allocation scheme. Indeed, resource alloca- tion is traditionally a Media Access Control (MAC) layer problem. This means that when solving a resource allocation problem in a network, the routing paths are usually assumed to be known. Conversely, the routing problem, which is a network layer issue, usually assumes that the available capacities on each link of the network (which depend on resource allocation) are known. Nevertheless, these two problems are mathematically entangled, and a cross-layer allocation strategy can best decoupled approaches in several ways, as we discuss in Chapter 4. Accordingly, our third and last contribution is to develop such a cross-layer allocation scheme for the scenario proposed in previous chapters.<p><p>All conclusions are summarised in Chapter 5, which also points to a few tracks for future research. / Doctorat en Sciences de l'ingénieur / info:eu-repo/semantics/nonPublished
3

Decision Strategy to Minimize Replenishment Costs in a Distribution Center with Forward Reserve Storage

Hollingsworth, Bradley K. 27 June 2003 (has links)
No description available.
4

Um modelo de seleção de carteiras de ações baseado em otimização convexa online

Yamim, João Daniel Madureira 23 February 2018 (has links)
Submitted by Geandra Rodrigues (geandrar@gmail.com) on 2018-05-23T12:11:21Z No. of bitstreams: 1 joaodanielmadureirayamim.pdf: 873324 bytes, checksum: 5025e3943c3bb2f1e1f19c55767c683e (MD5) / Approved for entry into archive by Adriana Oliveira (adriana.oliveira@ufjf.edu.br) on 2018-05-24T17:47:46Z (GMT) No. of bitstreams: 1 joaodanielmadureirayamim.pdf: 873324 bytes, checksum: 5025e3943c3bb2f1e1f19c55767c683e (MD5) / Made available in DSpace on 2018-05-24T17:47:46Z (GMT). No. of bitstreams: 1 joaodanielmadureirayamim.pdf: 873324 bytes, checksum: 5025e3943c3bb2f1e1f19c55767c683e (MD5) Previous issue date: 2018-02-23 / CAPES - Coordenação de Aperfeiçoamento de Pessoal de Nível Superior / Desde o trabalho seminal de Harry Markowitz, em 1952, que iniciou a moderna te-oria de carteiras, as estratégias de alocação de portfólio foram intensamente discutidas na literatura. Com o desenvolvimento de técnicas de otimização online, os algoritmos de aprendizado dinâmico se mostraram uma abordagem efetiva para construir portfólios (COVER, 1991; ARGAWAL et al., 2006). No entanto, poucos trabalhos conectam a lite-ratura tradicional, evoluída a partir do trabalho de Markowitz (1952) com a literatura de otimização online, que evoluiu a partir do trabalho de Cover (1991). O principal objetivo deste trabalho é implementar técnicas de otimização convexa online para: (i) executar estratégias de alocação de portfólio; (ii) conectar esses algoritmos com fatores risco usados em metodologias tradicionais. Dois métodos de algoritmos online foram implementados e adaptados, o Online Gradient Descendent (OGD) e o Online Newton Step (ONS). Além disso, duas novas versões para o algoritmo OGD são propostas para controlar o risco em carteiras. O primeiro, busca limitar o investimento máximo para ações e, o segundo, visa controlar o /3 das carteiras. Ambas as estratégias foram comparadas com o Uniform Constant Rebalanced Portfolio (UCRP) e o Dow Jones Industrial Index (DJIA). Foram utilizados dados do DJIA de março de 1987 até fevereiro de 2009 com observações se-manais. O algoritmo OGD apresentou o maior retorno acumulado entre as estratégias testadas. Ambos os algoritmos (OGD e ONS) apresentaram melhor desempenho do que o UCRP e DJIA ao longo do período. Além disso, o mecanismo de controle de risco pro-posto provou ser uma ferramenta útil para melhorar os resultados relacionados ao valor em risco (VaR) e ao valor condicional em risco (CVaR) das carteiras. / Since the seminal work of Harry Markowitz (1952), which initiated the modern theory of portfolios, the strategies of portfolio allocation were extensively discussed in economic literature. With the development of online optimization techniques, dynamic learning algorithms emerged as an effective approach to develop investment portfolios (COVER, 1991; ARGAWAL et al., 2006). However, there are few attempts aiming to connect the traditional literature of portfolio investment, which evolved based on Markowitz (1952) work, with the recent online methods, developed from Cover (1991). The main objec-tive of this work is to implement online convex optimization techniques to: (i) perform strategies of portfolio allocation; (ii) couple these algorithms with risk factors used in traditional models. Two methods of online algorithms were implemented and adapted, the Online Gradient Descendent (OGD) and the Online Newton Step (ONS). Besides, two new versions for the OGD algorithm are proposed in order to control risk in portfolios. The first one, seeks to limit maximum investment for stocks and, the second, aims to keep control of the /3 of portfolios. Both strategies were compared with the Uniform Constant Re-Balanced Portfolio (UCRP) and the Dow Jones Industrial Index (DJIA). Data from weekly observations of DJIA from March 1987 until February 2009 are used. The OGD algorithm presented the best accumulated return among all strategies. Both algorithms (OGD and ONS) performed better than the UCRP and DJIA index. Furthermore, the risk control mechanism proposed proved to be an useful tool in order to improve results related to the Value at Risk (VaR) and Conditional Value at Risk (CVaR) of the portfolios.
5

Resgate da otimalidade de estratégias de alocação dinâmica com seguro e alavancagem em cenários realistas

Varanda, José Henrique de Oliveira 02 July 2018 (has links)
Submitted by Sara Ribeiro (sara.ribeiro@ucb.br) on 2018-11-09T18:21:24Z No. of bitstreams: 1 JoseHenriquedeOliveiraVarandaDissertacao2018.pdf: 3107527 bytes, checksum: ea06abcabf1c014758cc880bcf0b0726 (MD5) / Approved for entry into archive by Sara Ribeiro (sara.ribeiro@ucb.br) on 2018-11-13T16:00:57Z (GMT) No. of bitstreams: 1 JoseHenriquedeOliveiraVarandaDissertacao2018.pdf: 3107527 bytes, checksum: ea06abcabf1c014758cc880bcf0b0726 (MD5) / Made available in DSpace on 2018-11-13T16:00:57Z (GMT). No. of bitstreams: 1 JoseHenriquedeOliveiraVarandaDissertacao2018.pdf: 3107527 bytes, checksum: ea06abcabf1c014758cc880bcf0b0726 (MD5) Previous issue date: 2018-07-02 / This study evaluates which modifications can restore the theoretical performance of dynamic asset allocation strategies that uses insurance and leverage, specifically those known as Constant Proportion Portfolio Insurance (CPPI), when confronted with realistic premises and scenarios. Simulations using GARCH models are applied to assess the effects of path dependency and volatility on those strategies and to evaluate how selected modifications mitigates those effects. These modifications are tested using the Farinelli- Tibilleti ratio and derivations, like de Upside Potential Ratio. As main finding, the modifications that mitigates path dependency can restore the theoretical performance of portfolio insurance with high significance, making those preferred strategies in relation to Buy-and-Hold (BH) or Constant-Mix (CM) for most investors in several scenarios. This work also presents a novel modification, adapted for the risk-free market in Brazil, that resulted in the best performing portfolio insurance strategy with great significance. / Este trabalho avalia quais modificações reestabelecem o desempenho teórico das estratégias dinâmicas de alocação de ativos com seguro e alavancagem, denominadas Constant Proportion Portfolio Insurance (CPPI), quando confrontadas com premissas e cenários realistas. São realizadas simulações de modelos da família GARCH, com parâmetros estimados do mercado, para exercitar os efeitos da dependência do caminho e da volatilidade nestas estratégias e avaliar como as modificações selecionadas ajudam a combate-los. A significância das modificações é testada pela medida Farinelli-Tibiletti, sobre tudo a combinação que resulta na razão Upside Potential, onde conclui-se que existem modificações significantes que são capazes de resgatar o desempenho teórico da estratégia CPPI, inclusive tornando-a preferível às estratégias clássicas Buy-and-Hold (BH) e Constant-Mix (CM) em certos cenários. Por fim, o trabalho apresenta uma modificação inovadora, derivada do ajuste à realidade do mercado brasileiro, que acabou por apresentar o maior nível de desempenho relativo do método CPPI, com elevada significância.
6

Essays on indexability of stochastic sheduling and dynamic allocation problems

Ruíz Hernández, Diego 13 April 2007 (has links)
In this Thesis, we first deploy Gittins index theory to establish the indexability of inter-alia general families of restless bandits that arise in problems of stochastic scheduling with switching penalties and machine maintenance. We also give formulae for the resulting indices. Numerical investigations testify the strong performance of the index heuristics.The second class of problems concerns two families of Markov decision problems. The spinning plates problem concerns the optimal management of a portfolio of assets whose yields grow with investment but otherwise decline. In the model of asset exploitation called the squad system, the yield from an asset declines when it is utilised but will recover when the asset is at rest. Simply stated conditions are given which guarantee general indexability of the problem together with necessary and sufficient conditions for strict indexability. The index heuristics, which emerge from the analysis, are assessed numerically and found to perform strongly.
7

Conception d'un "front-end" RF millimétrique pour un système de communication sur puce multi-accès innovant utilisant un réseau d'interconnexions RF-NoC / Wired RF-based Network On Chip Reconfigurable On Demand

Drillet, Frédéric 14 October 2016 (has links)
Résumé des travaux de thèse Frédéric DRILLETThèse intitulée : Conception d'un front-end RF de bande passante [20-40] GHz pour un système de communication sur puce utilisant un réseau d'interconnexions RF-NoC.Technologie : NXP Qubic4XI (BiCMOS SiGe:C 250 nm)Résumé :La tendance actuelle dans la conception de systèmes sur puces (SoC) est d'intégrer un très grand nombre d'unités de calcul et de mémoires sur une seule puce. Les possibilités de cette intégration poussée permettent aujourd'hui d'envisager le développement d'une électronique offrant une multitude de services. Néanmoins ces architectures posent de nouveaux challenges concernant les interconnexions entre les unités de calcul. En effet, pour les futures générations technologiques, la mise à l'échelle impactera lourdement les performances des interconnexions globales en termes de débit, latence et consommation. Afin de répondre à la problématique des communications intra-puces, un certain nombre de technologies ont été investiguées comme les technologies d’intégration 3D, les architectures utilisant l'optique ou la RF. L'approche RF pour les communications entre les unités de calcul d’un même circuit de type NoC (Network On Chip) présente l'avantage d'une bonne compatibilité avec les technologies CMOS silicium et peut également répondre aux besoins de communication dans les structures 3D.Cette thèse s'inclue dans le projet ANR WiNoCoD qui propose un réseau d'interconnexion RF-NoC utilisant l'OFDMA. Elle porte sur la conception d'un front-end RF générique permettant de transmettre et de recevoir toute la bande passante soit [20-40] GHz. Cette généricité permet une allocation dynamique des porteuses sans reconfiguration du hardware. On utilise la technologie QubiC4XI de NXP Semiconductors, qui est une technologie BiCMOS SiGe:C 250 nm, afin de vérifier la faisabilité d'un tel système avec des moyens actuels. Ce front-end doit être large bande puisqu'il a une bande passante de 20 GHz entre 20 et 40 GHz. Il doit également consommer le moins possible puisqu'il a pour but d'être intégré dans un système contenant plusieurs NoC et qui est donc très énergivore. Il doit également être compact pour ne pas occuper plus de surface que la partie numérique.Cette thèse inclue la conception des éléments composant le front-end, les résultats de simulation et de mesure, ainsi que les performances du système complet. / Frédéric DRILLET thesis work summaryThesis entitled: Design of a [20-40] GHz RF front-end for an on-chip RF-NoC communication system.Technology: NXP Qubic4XI (BiCMOS SiGe:C 250 nm)Summary:A current trend regarding System On Chip design is to include a very big amount of processors and memories on a single chip. Today, these integrated circuits allow to consider an electronic supplying a multitude of services. However, these architectures are challenging in terms of connection between processing units. It could indeed lead to data rate, latency and consumption degradation. In order to overcome these issues technological solutions were investigated such as 3D integration, or optic and RF networks. An RF Network on Chip (NoC) is compatible with silicon CMOS technologies and with 3D structures.This thesis is a part of the ANR project called WiNoCoD (Wired Network on Chip reconfigurable on Demand) which offers an OFDMA RF-NoC. The main work presents a generic RF front-end allowing to transmit and receive the whole [20-40] GHz bandwidth. This generic architecture allows a dynamic allocation of OFDMA subcarriers without any hardware reconfiguration. The technology used is the NXP Semiconductor QubiC4XI which is a BiCMOS SiGe:C 250 nm technology. A current technology is used to check the feasibility of such a system today. This front-end has to be wideband. The power consumption has to be as low as possible as well, as it is going to be integrated in a system containing several NoCs that consume already a lot of power. The system has to be very compact, its total area has to be smaller than the digital part.This thesis includes the design of the front end, the simulation and measurement results and the performance of the full system.
8

Hybrid real-time operating system integrated with middleware for resource-constrained wireless sensor nodes / Système d'exploitation temps-réel hybride intégré avec un middelware pour les noeuds capteurs sans fil contraints en ressources

Liu, Xing 30 June 2014 (has links)
Avec les avancées récentes en microélectronique, en traitement numérique et en technologie de communication, les noeuds de réseau de capteurs sans fil (noeud RCSF) deviennent de moins en moins encombrants et coûteux. De ce fait la technologie de RCSF est utilisée dans de larges domaines d’application. Comme les noeuds RCSF sont limités en taille et en coût, ils sont en général équipés d’un petit microcontrôleur de faible puissance de calcul et de mémoire etc. De plus ils sont alimentés par une batterie donc son énergie disponible est limitée. A cause de ces contraintes, la plateforme logicielle d’un RCSF doit consommer peu de mémoire, d’énergie, et doit être efficace en calcul. Toutes ces contraintes rendent les développements de logiciels dédiés au RCSF très compliqués. Aujourd’hui le développement d’un système d’exploitation dédié à la technologie RCSF est un sujet important. En effet avec un système d’exploitation efficient, les ressources matérielles d’une plateforme RCSF peuvent être utilisées efficacement. De plus, un ensemble de services système disponibles permet de simplifier le développement d’une application. Actuellement beaucoup de travaux de recherche ont été menés pour développer des systèmes d’exploitation pour le RCSF tels que TinyOS, Contiki, SOS, openWSN, mantisOS et simpleRTJ. Cependant plusieurs défis restent à relever dans le domaine de système d’exploitation pour le RCSF. Le premier des défis est le développement d’un système d’exploitation temps réel à faible empreinte mémoire dédié au RCSF. Le second défi est de développer un mécanisme permettant d’utiliser efficacement la mémoire et l’énergie disponible d’un RCSF. De plus, comment fournir un développement d’application pour le RCSF reste une question ouverte. Dans cette thèse, un nouveau système d’exploitation hybride, temps réel à énergie efficiente et à faible empreinte mémoire nommé MIROS dédié au RCSF a été développé. Dans MIROS, un ordonnanceur hybride a été adopté ; les deux ordonnanceurs évènementiel et multithread ont été implémentés. Avec cet ordonnanceur hybride, le nombre de threads de MIROS peut être diminué d’une façon importante. En conséquence, les avantages d’un système d’exploitation évènementiel qui consomme peu de ressource mémoire et la performance temps réel d’un système d’exploitation multithread ont été obtenues. De plus, l’allocation dynamique de la mémoire a été aussi réalisée dans MIROS. La technique d’allocation mémoire de MIROS permet l’augmentation de la zone mémoire allouée et le réassemblage des fragments de mémoire. De ce fait, l’allocation de mémoire de MIROS devient plus flexible et la ressource mémoire d’un noeud RCSF peut être utilisée efficacement. Comme l’énergie d’un noeud RCSF est une ressource à forte contrainte, le mécanisme de conservation d’énergie a été implanté dans MIROS. Contrairement aux autres systèmes d’exploitation pour RCSF où la conservation d’énergie a été prise en compte seulement en logiciel, dans MIROS la conservation d’énergie a été prise en compte à la fois en logiciel et en matériel. Enfin, pour fournir un environnement de développement convivial aux utilisateurs, un nouveau intergiciel nommé EMIDE a été développé et intégré dans MIROS. EMIDE permet le découplage d’une application de système. Donc le programme d’application est plus simple et la reprogrammation à distance est plus performante, car seulement les codes de l’application seront reprogrammés. Les évaluations de performance de MIROS montrent que MIROS est un système temps réel à faible empreinte mémoire et efficace pour son exécution. De ce fait, MIROS peut être utilisé dans plusieurs plateformes telles que BTnode, IMote, SenseNode, TelosB et T-Mote Sky. Enfin, MIROS peut être utilisé pour les plateformes RCSF à fortes contraintes de ressources. / With the recent advances in microelectronic, computing and communication technologies, wireless sensor network (WSN) nodes have become physically smaller and more inexpensive. As a result, WSN technology has become increasingly popular in widespread application domains. Since WSN nodes are minimized in physical size and cost, they are mostly restricted to platform resources such as processor computation ability, memory resources and energy supply. The constrained platform resources and diverse application requirements make software development on the WSN platform complicated. On the one hand, the software running on the WSN platform should be small in the memory footprint, low in energy consumption and high in execution efficiency. On the other hand, the diverse application development requirements, such as the real-time guarantee and the high reprogramming performance, should be met by the WSN software. The operating system (OS) technology is significant for the WSN proliferation. An outstanding WSN OS can not only utilize the constrained WSN platform resources efficiently, but also serve the WSN applications soundly. Currently, a set of WSN OSes have been developed, such as the TinyOS, the Contiki, the SOS, the openWSN and the mantisOS. However, many OS development challenges still exist, such as the development of a WSN OS which is high in real-time performance yet low in memory footprint; the improvement of the utilization efficiency to the memory and energy resources on the WSN platforms, and the providing of a user-friendly application development environment to the WSN users. In this thesis, a new hybrid, real-time, energy-efficient, memory-efficient, fault-tolerant and user-friendly WSN OS MIROS is developed. MIROS uses the hybrid scheduling to combine the advantages of the event-driven system's low memory consumption and the multithreaded system's high real-time performance. By so doing, the real-time scheduling can be achieved on the severely resource-constrained WSN platforms. In addition to the hybrid scheduling, the dynamic memory allocators are also realized in MIROS. Differing from the other dynamic allocation approaches, the memory heap in MIROS can be extended and the memory fragments in the MIROS can be defragmented. As a result, MIROS allocators become flexible and the memory resources can be utilized more efficiently. Besides the above mechanisms, the energy conservation mechanism is also implemented in MIROS. Different from most other WSN OSes in which the energy resource is conserved only from the software aspect, the energy conservation in MIROS is achieved from both the software aspect and the multi-core hardware aspect. With this conservation mechanism, the energy cost reduced significantly, and the lifetime of the WSN nodes prolonged. Furthermore, MIROS implements the new middleware software EMIDE in order to provide a user-friendly application development environment to the WSN users. With EMIDE, the WSN application space can be decoupled from the low-level system space. Consequently, the application programming can be simplified as the users only need to focus on the application space. Moreover, the application reprogramming performance can be improved as only the application image other than the monolithic image needs to be updated during the reprogramming process. The performance evaluation works to the MIROS prove that MIROS is a real-time OS which has small memory footprint, low energy cost and high execution efficiency. Thus, it is suitable to be used on many WSN platforms including the BTnode, IMote, SenseNode, TelosB, T-Mote Sky, etc. The performance evaluation to EMIDE proves that EMIDE has less memory cost and low energy consumption. Moreover, it supports small-size application code. Therefore, it can be used on the high resource-constrained WSN platforms to provide a user-friendly development environment to the WSN users.
9

Projeto e avaliação de um broker como agente de intermediação e QoS em uma nuvem computacional híbrida / Design and evaluation of a broker as QoS and intermediation agent in hybrid cloud computing

Pardo, Mario Henrique de Souza 16 June 2016 (has links)
A presente tese de doutorado propõe uma arquitetura de cloud broker para ambientes de computação em nuvem híbrida. Um cloud broker tem o objetivo de executar a mediação entre clientes e provedores, recebendo requisições dos clientes e encaminhando-as ao serviço do provedor que melhor se adaptar aos requisitos de qualidade de serviço (QoS) solicitados. A arquitetura de broker de serviços com QoS proposta denomina-se QBroker, características de implementação de seu modo de operação bem como sua interação com os recursos virtuais de um ambiente de nuvem são apresentadas. O modelo de nuvem considerado foi o de nuvem híbrida com uma caracterização de arquitetura orientada a serviços (SOA) na qual serviços remotos são disponibilizados aos clientes. A política de escalonamento de tarefas desenvolvida para o QBroker foi a de intermediação de serviços, considerando tratativas de QoS, diferenciação das instâncias de serviços (SOA) e alocação dinâmica de serviços. Além disso, toda a caracterização do modo de operação do QBroker foi baseada no conceito de intermediação do modelo de referência de nuvem do NIST. O componente QBroker foi introduzido numa arquitetura de computação em nuvem BEQoS (Bursty Energy and Quality of Service), desenvolvida no Laboratório de Sistemas Distribuídos e Programação Concorrente do ICMC-USP de São Carlos. Avaliações de desempenho para a implementação da arquitetura QBroker foram conduzidas por meio de programas de simulação com uso da API do simulador CloudSim e da arquitetura CloudSim-BEQoS. Três cenários experimentais foram avaliados e, segundo a análise de resultados efetuada, foi possível validar que as características arquiteturais implementadas no QBroker resultaram em significativo impacto nas variáveis de resposta consideradas. Assim, foi possível comprovar que o uso do QBroker como mecanismo de mediação em ambientes de nuvem híbrida com SOA promoveu ganhos em desempenho para o sistema de nuvem e permitiu melhoria na qualidade dos serviços oferecidos. / This doctoral thesis proposes a cloud broker architecture for hybrid cloud computing environments. A cloud broker aims to perform mediation between clients and providers, receiving customer requests and forwarding them to the service provider that best suits the requested QoS requirements. The broker architecture services with QoS proposal is called QBroker. Implementation features of its mode of operation as well as its interaction with the virtual resources from a cloud environment are presented. The cloud deployment model was considered a hybrid cloud with a characterization of service-oriented architecture (SOA) in which remote services are available to customers. The task scheduling policy developed for QBroker was the intermediation of services, considering negotiations of QoS, differentiation of services instances and dynamic allocation of services. Moreover, the entire characterization of QBroker operation mode is based on the intermediation concept of the NIST cloud reference model. The QBroker component was introduced into a cloud computing architecture BEQoS (Bursty, Energy and Quality of Service), developed in the Laboratory of Distributed Systems and Concurrent Programming at ICMC-USP. Performance evaluations analysis the of results of QBroker architecture were conducted through simulation programs using the CloudSim simulator API and CloudSim-BEQoS architecture. Three experimental scenarios were evaluated and, according to analysis of the results, it was possible to validate that the architectural features implemented in QBroker resulted in significant impact on response variables considered. Thus, it was possible to prove that the use of QBroker as mediation mechanism in hybrid cloud environments with SOA promoted performance gains for the cloud system and allowed improvement in the quality of services offered.
10

Projeto e avaliação de um broker como agente de intermediação e QoS em uma nuvem computacional híbrida / Design and evaluation of a broker as QoS and intermediation agent in hybrid cloud computing

Mario Henrique de Souza Pardo 16 June 2016 (has links)
A presente tese de doutorado propõe uma arquitetura de cloud broker para ambientes de computação em nuvem híbrida. Um cloud broker tem o objetivo de executar a mediação entre clientes e provedores, recebendo requisições dos clientes e encaminhando-as ao serviço do provedor que melhor se adaptar aos requisitos de qualidade de serviço (QoS) solicitados. A arquitetura de broker de serviços com QoS proposta denomina-se QBroker, características de implementação de seu modo de operação bem como sua interação com os recursos virtuais de um ambiente de nuvem são apresentadas. O modelo de nuvem considerado foi o de nuvem híbrida com uma caracterização de arquitetura orientada a serviços (SOA) na qual serviços remotos são disponibilizados aos clientes. A política de escalonamento de tarefas desenvolvida para o QBroker foi a de intermediação de serviços, considerando tratativas de QoS, diferenciação das instâncias de serviços (SOA) e alocação dinâmica de serviços. Além disso, toda a caracterização do modo de operação do QBroker foi baseada no conceito de intermediação do modelo de referência de nuvem do NIST. O componente QBroker foi introduzido numa arquitetura de computação em nuvem BEQoS (Bursty Energy and Quality of Service), desenvolvida no Laboratório de Sistemas Distribuídos e Programação Concorrente do ICMC-USP de São Carlos. Avaliações de desempenho para a implementação da arquitetura QBroker foram conduzidas por meio de programas de simulação com uso da API do simulador CloudSim e da arquitetura CloudSim-BEQoS. Três cenários experimentais foram avaliados e, segundo a análise de resultados efetuada, foi possível validar que as características arquiteturais implementadas no QBroker resultaram em significativo impacto nas variáveis de resposta consideradas. Assim, foi possível comprovar que o uso do QBroker como mecanismo de mediação em ambientes de nuvem híbrida com SOA promoveu ganhos em desempenho para o sistema de nuvem e permitiu melhoria na qualidade dos serviços oferecidos. / This doctoral thesis proposes a cloud broker architecture for hybrid cloud computing environments. A cloud broker aims to perform mediation between clients and providers, receiving customer requests and forwarding them to the service provider that best suits the requested QoS requirements. The broker architecture services with QoS proposal is called QBroker. Implementation features of its mode of operation as well as its interaction with the virtual resources from a cloud environment are presented. The cloud deployment model was considered a hybrid cloud with a characterization of service-oriented architecture (SOA) in which remote services are available to customers. The task scheduling policy developed for QBroker was the intermediation of services, considering negotiations of QoS, differentiation of services instances and dynamic allocation of services. Moreover, the entire characterization of QBroker operation mode is based on the intermediation concept of the NIST cloud reference model. The QBroker component was introduced into a cloud computing architecture BEQoS (Bursty, Energy and Quality of Service), developed in the Laboratory of Distributed Systems and Concurrent Programming at ICMC-USP. Performance evaluations analysis the of results of QBroker architecture were conducted through simulation programs using the CloudSim simulator API and CloudSim-BEQoS architecture. Three experimental scenarios were evaluated and, according to analysis of the results, it was possible to validate that the architectural features implemented in QBroker resulted in significant impact on response variables considered. Thus, it was possible to prove that the use of QBroker as mediation mechanism in hybrid cloud environments with SOA promoted performance gains for the cloud system and allowed improvement in the quality of services offered.

Page generated in 0.5136 seconds