• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 85
  • 11
  • 9
  • 7
  • 6
  • 6
  • 2
  • 2
  • 2
  • 2
  • 1
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 162
  • 162
  • 34
  • 33
  • 30
  • 28
  • 27
  • 24
  • 23
  • 20
  • 20
  • 19
  • 18
  • 18
  • 17
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
111

VM allocation in cloud datacenters based on the multi-agent system : an investigation into the design and response time analysis of a multi-agent-based virtual machine (VM) allocation/placement policy in cloud datacenters

Al-ou'n, Ashraf M. S. January 2017 (has links)
Recent years have witnessed a surge in demand for infrastructure and services to cover high demands on processing big chunks of data and applications resulting in a mega Cloud Datacenter. A datacenter is of high complexity with increasing difficulties to identify, allocate efficiently and fast an appropriate host for the requested virtual machine (VM). Establishing a good awareness of all datacenter’s resources enables the allocation “placement” policies to make the best decision in reducing the time that is needed to allocate and create the VM(s) at the appropriate host(s). However, current algorithms and policies of placement “allocation” do not focus efficiently on awareness of the resources of the datacenter, and moreover, they are based on conventional static techniques. Which are adversely impacting on the allocation progress of the policies. This thesis proposes a new Agent-based allocation/placement policy that employs some of the Multi-Agent system features to get a good awareness of Cloud Datacenter resources and also provide an efficient allocation decision for the requested VMs. Specifically, (a) The Multi-Agent concept is used as a part of the placement policy (b) A Contract Net Protocol is devised to establish good awareness and (c) A verification process is developed to fully dimensional VM specifications during allocation. These new results show a reduction in response time of VM allocation and the usage improvement of occupied resources. The proposed Agent-based policy was implemented using the CloudSim toolkit and consequently was compared, based on a series of typical numerical experiments, with the toolkit’s default policy. The comparative study was carried out in terms of the time duration of VM allocation and other aspects such as the number of available VM types and the amount of occupied resources. Moreover, a two-stage comparative study was introduced through this thesis. Firstly, the proposed policy is compared with four state of the art algorithms, namely the Random algorithm and three one-dimensional Bin-Packing algorithms. Secondly, the three Bin-Packing algorithms were enhanced to have a two-dimensional verification structure and were compared against the proposed new algorithm of the Agent-based policy. Following a rigorous comparative study, it was shown that, through the typical numerical experiments of all stages, the proposed new Agent-based policy had superior performance in terms of the allocation times. Finally, avenues arising from this thesis are included.
112

Optimization of routing and wireless resource allocation in hybrid data center networks / Optimisation du routage et de l'allocation de ressources sans fil dans les réseaux des centres de données hybrides

Dab, Boutheina 05 July 2017 (has links)
L’arrivée de la prochaine technologie 5G va permettre la connectivité des billions de terminaux mobiles et donc une énorme augmentation du trafic de données. A cet égard, les fournisseurs des services Cloud doivent posséder les infrastructures physiques capables de supporter cette explosion de trafic. Malheureusement, les architectures filaires conventionnelles des centres de données deviennent staturées et la congestion des équipements d’interconnexion est souvent atteinte. Dans cette thèse, nous explorons une approche récente qui consiste à augmenter le réseau filaire du centre de données avec l’infrastructure sans fil. En effet, nous exploitons une nouvelle technologie émergente, la technologie 60 GHz, qui assure un débit de l’ordre de 7 Gbits/s afin d’améliorer la QoS. Nous concevons une architecture hybride (filaire/sans fil) du réseau de centre de données basée sur : i) le modèle "Cisco’s Massively Scalable Data Center" (MSDC), et ii) le standard IEEE 802.11ad. Dans une telle architecture, les serveurs sont regroupés dans des racks, et sont interconnectés à travers un switch Ethernet, appelé top-of-rack (ToR) switch. Chaque ToR switch possède plusieurs antennes utilisées en parallèle sur différents canaux sans fil. L’objectif final consiste à minimiser la congestion du réseau filaire, en acheminant le maximum du trafic sur les canaux sans fil. Pour ce faire, cette thèse se focalise sur l’optimisation du routage et de l’allocation des canaux sans fil pour les communications inter-rack, au sein d’un centre de données hybride (HDCN). Ce problème étant NP-difficile, nous allons procéder en trois étapes. En premier lieu, on considère le cas des communications à un saut, où les racks sont placés dans le même rayon de transmission. Nous proposons un nouvel algorithme d’allocation des canaux sans fil dans les HDCN, qui permet d’acheminer le maximum des communications en sans-fil, tout en améliorant les performances réseau en termes de débit et délai. En second lieu, nous nous adressons au cas des communications à plusieurs sauts, où les racks ne sont pas dans le même rayon de transmission. Nous allons proposer une nouvelle approche optimale traitant conjointement le problème du routage et de l’allocation de canaux sans fils dans le HDCN, pour chaque communication, dans un mode online. En troisième étape, nous proposons un nouvel algorithme qui calcule conjointement le routage et l’allocation des canaux pour un ensemble des communications arrivant en mode batch (i.e., par lot). En utilisant le simulateur réseau QualNet, considérant toute la pile TCP/IP, les résultats obtenus montrent que nos propositions améliorent les performances comparées aux méthodes de l’état de l’art / The high proliferation of smart devices and online services allows billions of users to connect with network while deploying a vast range of applications. Particularly, with the advent of the future 5G technology, it is expected that a tremendous mobile and data traffic will be crossing Internet network. In this regard, Cloud service providers are urged to rethink their data center architectures in order to cope with this unprecedented traffic explosion. Unfortunately, the conventional wired infrastructures struggle to resist to such a traffic growth and become prone to serious congestion problems. Therefore, new innovative techniques are required. In this thesis, we investigate a recent promising approach that augments the wired Data Center Network (DCN) with wireless communications. Indeed, motivated by the feasibility of the new emerging 60 GHz technology, offering an impressive data rate (≈ 7 Gbps), we envision, a Hybrid (wireless/wired) DCN (HDCN) architecture. Our HDCN is based on i) Cisco’s Massively Scalable Data Center (MSDC) model and ii) IEEE 802.11ad standard. Servers in the HDCN are regrouped into racks, where each rack is equipped with a: i) Ethernet top-of-rack (ToR) switch and ii) set of wireless antennas. Our research aims to optimize the routing and the allocation of wireless resources for inter-rack communications in HDCN while enhancing network performance and minimizing congestion. The problem of routing and resource allocation in HDCN is NP-hard. To deal with this difficulty, we will tackle the problem into three stages. In the first stage, we consider only one-hop inter-rack communications in HDCN, where all communicating racks are in the same transmission range. We will propound a new wireless channel allocation approach in HDCN to hardness both wireless and wired interfaces for incoming flows while enhancing network throughput. In the second stage, we deal with the multi-hop communications in HDCN where communicating racks can not communicate in one single-hop wireless path. We propose a new approach to jointly route and allocate channels for each single communication flow, in an online way. Finally, in the third stage, we address the batched arrival of inter-rack communications to the HDCN so as to further optimize the usage of wireless and wired resources. For that end, we propose: i) a heuristic-based and ii) an approximate, solutions, to solve the joint batch routing and channel assignment. Based on extensive simulations conducted in QualNet simulator while considering the full protocol stack, the obtained results for both real workload and uniform traces, show that our proposals outperform the prominent related strategies
113

Mitteilungen des URZ 3/2009

Clauß, Matthias, Müller, Thomas, Riedel, Wolfgang, Schier, Thomas, Vodel, Matthias 31 August 2009 (has links)
Informationen des Universitätsrechenzentrums
114

Sustainable Resource Management for Cloud Data Centers

Mahmud, A. S. M. Hasan 15 June 2016 (has links)
In recent years, the demand for data center computing has increased significantly due to the growing popularity of cloud applications and Internet-based services. Today's large data centers host hundreds of thousands of servers and the peak power rating of a single data center may even exceed 100MW. The combined electricity consumption of global data centers accounts for about 3% of worldwide production, raising serious concerns about their carbon footprint. The utility providers and governments are consistently pressuring data center operators to reduce their carbon footprint and energy consumption. While these operators (e.g., Apple, Facebook, and Google) have taken steps to reduce their carbon footprints (e.g., by installing on-site/off-site renewable energy facility), they are aggressively looking for new approaches that do not require expensive hardware installation or modification. This dissertation focuses on developing algorithms and systems to improve the sustainability in data centers without incurring significant additional operational or setup costs. In the first part, we propose a provably-efficient resource management solution for a self-managed data center to cap and reduce the carbon emission while maintaining satisfactory service performance. Our solution reduces the carbon emission of a self-managed data center to net-zero level and achieves carbon neutrality. In the second part, we consider minimizing the carbon emission in a hybrid data center infrastructure that includes geographically distributed self-managed and colocation data centers. This segment identifies and addresses the challenges of resource management in a hybrid data center infrastructure and proposes an efficient distributed solution to optimize the workload and resource allocation jointly in both self-managed and colocation data centers. In the final part, we explore sustainable resource management from cloud service users' point of view. A cloud service user purchases computing resources (e.g., virtual machines) from the service provider and does not have direct control over the carbon emission of the service provider's data center. Our proposed solution encourages a user to take part in sustainable (both economical and environmental) computing by limiting its spending on cloud resource purchase while satisfying its application performance requirements.
115

Cloud computing s ohledem na technologické aspekty a změny v infrastruktuře / Cloud computing with regard to technological aspects and changes in infrastructure

Vítek, Daniel January 2010 (has links)
This thesis discusses the new way of delivering IT services over the Internet widely known as cloud computing. In its opening part, cloud computing is put into a historical context of the evolution of enterprise computing, and the dominant issues the IT department faces today are mentioned. Further, the paper deals with several components that make up the architecture of cloud computing and reviews the benefits and drawbacks an enterprise can have while it adopts this new model. One of the primary aims of this thesis is to identify the impact of the technology trends on cloud computing. The thesis brings together four major computing trends, namely virtualization, multi-tenant architecture, service-oriented architecture and grid computing. Another aim is to focus on two trends related to IT infrastructure that will lead to fundamental changes in IT industry. The first of them is the emergence of extremely large-scale data centers at low cost locations, which can serve tremendous amount of customers and achieve considerable economies of scale. The second trend this paper points out is the shift from multi-purpose all-in-one computers into a wide range of mobile devices dedicated to a specific user's needs. The last aim of this thesis is to clarify the economic impact of cloud computing in terms of costs and changes in business models. The thesis concludes by evaluating the current adoption and predicting the future trend of cloud computing.
116

A Component-based Business Continuity and Disaster Recovery Framework

Somasekaram, Premathas January 2017 (has links)
IT solutions must be protected so that the business can continue, even in the case of fatal failures associated with disasters. Business continuity in the context of disaster implies that business cannot continue in the current environment but instead must continue at an alternate site or data center. However, the BC/DR concept today is too fragmented, as many different frameworks and methodologies exist. Furthermore,many of the application-specific solutions are provided and promoted by software vendors, while hardware vendors provide solutions for their hardware environments. Nevertheless, there are concerns that BC/DR solutions often do not connect to the technical components that are in the lower layers, which function as the foundationfor any such solutions; hence, it is equally important to connect and map the requirements accordingly. Moreover, a shift in the hardware environment, such as cloud computing, as well as changes in operations management, such as outsourcing,add complexity that must be captured by a BC/DR solution. Furthermore, the integrated nature of IT-based business solutions also presents new challenges, as it isno longer one IT solution that must be protected but also other IT solutions that are integrated to deliver an individual business process. Thus, it will be difficult to employa current BC/DR approach. Hence, the purpose of this thesis project is to design, develop, and present a novel way of addressing the BC/DR gaps, while supporting the requirements of a dynamic IT environment. The solution reuses most elements fromthe existing standards and solutions. However, it also includes new elements to capture and present the technical solution; hence, the complete solution is designatedas a framework. The new framework can support many IT solutions since it will havea modular approach, and it is flexible, scalable, and platform and application independent, while addressing the solution on a component level. The new framework is applied to two application scenarios at the stakeholder site, and theresults are studied and presented in this thesis.
117

Cenové modely a SLA pre zdieľané služby dátových centier / Price models and SLA for shared services of data centers

Štefek, Ivan January 2012 (has links)
Thesis Price models and SLA for shared services of data centers has as objective to establish best practice, so-called "best practices" of SLA and price models for data center services. Work is divided into two parts, theoretical and practical part. The theoretical section is wants to familiarize the reader with the basic concepts such as outsourcing, reasons and risks of use. I describe the base models for providing IT services as well as pricing models for services as described in the literature. I will focus on data centers and TIER standard categorization, too. The practical part is based on an analysis of information that I got from an interview with the experts working for the biggest data center companies in the Czech Republic. Based on this information, I tried to compile the best practices for creating pricing models for data center services that are actually used in practice. Specifically, the server housing, server hosting and virtual server services. Also, for these services create an SLA that can be used by some data center company in the Czech Republic.
118

Efficient and elastic management of computing infrastructures

Alfonso Laguna, Carlos de 23 October 2016 (has links)
[EN] Modern data centers integrate a lot of computer and electronic devices. However, some reports state that the mean usage of a typical data center is around 50% of its peak capacity, and the mean usage of each server is between 10% and 50%. A lot of energy is destined to power on computer hardware that most of the time remains idle. Therefore, it would be possible to save energy simply by powering off those parts from the data center that are not actually used, and powering them on again as they are needed. Most data centers have computing clusters that are used for intensive computing, recently evolving towards an on-premises Cloud service model. Despite the use of low consuming components, higher energy savings can be achieved by dynamically adapting the system to the actual workload. The main approach in this case is the usage of energy saving criteria for scheduling the jobs or the virtual machines into the working nodes. The aim is to power off idle servers automatically. But it is necessary to schedule the power management of the servers in order to minimize the impact on the end users and their applications. The objective of this thesis is the elastic and efficient management of cluster infrastructures, with the aim of reducing the costs associated to idle components. This objective is addressed by automating the power management of the working nodes in a computing cluster, and also proactive stimulating the load distribution to achieve idle resources that could be powered off by means of memory overcommitment and live migration of virtual machines. Moreover, this automation is of interest for virtual clusters, as they also suffer from the same problems. While in physical clusters idle working nodes waste energy, in the case of virtual clusters that are built from virtual machines, the idle working nodes can waste money in commercial Clouds or computational resources in an on-premises Cloud. / [ES] En los Centros de Procesos de Datos (CPD) existe una gran concentración de dispositivos informáticos y de equipamiento electrónico. Sin embargo, algunos estudios han mostrado que la utilización media de los CPD está en torno al 50%, y que la utilización media de los servidores se encuentra entre el 10% y el 50%. Estos datos evidencian que existe una gran cantidad de energía destinada a alimentar equipamiento ocioso, y que podríamos conseguir un ahorro energético simplemente apagando los componentes que no se estén utilizando. En muchos CPD suele haber clusters de computadores que se utilizan para computación de altas prestaciones y para la creación de Clouds privados. Si bien se ha tratado de ahorrar energía utilizando componentes de bajo consumo, también es posible conseguirlo adaptando los sistemas a la carga de trabajo en cada momento. En los últimos años han surgido trabajos que investigan la aplicación de criterios energéticos a la hora de seleccionar en qué servidor, de entre los que forman un cluster, se debe ejecutar un trabajo o alojar una máquina virtual. En muchos casos se trata de conseguir equipos ociosos que puedan ser apagados, pero habitualmente se asume que dicho apagado se hace de forma automática, y que los equipos se encienden de nuevo cuando son necesarios. Sin embargo, es necesario hacer una planificación de encendido y apagado de máquinas para minimizar el impacto en el usuario final. En esta tesis nos planteamos la gestión elástica y eficiente de infrastructuras de cálculo tipo cluster, con el objetivo de reducir los costes asociados a los componentes ociosos. Para abordar este problema nos planteamos la automatización del encendido y apagado de máquinas en los clusters, así como la aplicación de técnicas de migración en vivo y de sobreaprovisionamiento de memoria para estimular la obtención de equipos ociosos que puedan ser apagados. Además, esta automatización es de interés para los clusters virtuales, puesto que también sufren el problema de los componentes ociosos, sólo que en este caso están compuestos por, en lugar de equipos físicos que gastan energía, por máquinas virtuales que gastan dinero en un proveedor Cloud comercial o recursos en un Cloud privado. / [CAT] En els Centres de Processament de Dades (CPD) hi ha una gran concentració de dispositius informàtics i d'equipament electrònic. No obstant això, alguns estudis han mostrat que la utilització mitjana dels CPD està entorn del 50%, i que la utilització mitjana dels servidors es troba entre el 10% i el 50%. Estes dades evidencien que hi ha una gran quantitat d'energia destinada a alimentar equipament ociós, i que podríem aconseguir un estalvi energètic simplement apagant els components que no s'estiguen utilitzant. En molts CPD sol haver-hi clusters de computadors que s'utilitzen per a computació d'altes prestacions i per a la creació de Clouds privats. Si bé s'ha tractat d'estalviar energia utilitzant components de baix consum, també és possible aconseguir-ho adaptant els sistemes a la càrrega de treball en cada moment. En els últims anys han sorgit treballs que investiguen l'aplicació de criteris energètics a l'hora de seleccionar en quin servidor, d'entre els que formen un cluster, s'ha d'executar un treball o allotjar una màquina virtual. En molts casos es tracta d'aconseguir equips ociosos que puguen ser apagats, però habitualment s'assumix que l'apagat es fa de forma automàtica, i que els equips s'encenen novament quan són necessaris. No obstant això, és necessari fer una planificació d'encesa i apagat de màquines per a minimitzar l'impacte en l'usuari final. En esta tesi ens plantegem la gestió elàstica i eficient d'infrastructuras de càlcul tipus cluster, amb l'objectiu de reduir els costos associats als components ociosos. Per a abordar este problema ens plantegem l'automatització de l'encesa i apagat de màquines en els clusters, així com l'aplicació de tècniques de migració en viu i de sobreaprovisionament de memòria per a estimular l'obtenció d'equips ociosos que puguen ser apagats. A més, esta automatització és d'interés per als clusters virtuals, ja que també patixen el problema dels components ociosos, encara que en este cas estan compostos per, en compte d'equips físics que gasten energia, per màquines virtuals que gasten diners en un proveïdor Cloud comercial o recursos en un Cloud privat. / Alfonso Laguna, CD. (2015). Efficient and elastic management of computing infrastructures [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/57187 / TESIS
119

Zátěžové testy non-IT částí datových center / Load tests of non-IT parts of data centers

Víteček, Aleš January 2019 (has links)
The thesis "Load testing of non-IT parts data centers" deals with the theoretical analysis of data centers with a focus on non-IT part and theoretical preparation for measuring stress tests focused mainly on cooling. The theory is followed by a practical part of the PEC measurement by CATEGORY a.s .. This sample measurement carried out at the company premises clearly outlines the importance of performing similar tests. The measurement is thoroughly evaluated and suggestions are made to improve the operation. The next part of the practical part contains simulation of sample measurement. This simulation serves as a convenient addition and comparison. Also this simulation, the importance of similar stress tests can be justified.
120

Projekt datového centra - strukturovaná kabeláž / Data centre project - structured cabling

Hejtmánek, Ondřej January 2012 (has links)
This master’s thesis is concerned with analysis of cabling and cabling systems in data center and with their projection as follows. The part of the work is aimed at the measurement parameters of metal cables which are used in data center with noise. The first part is focused on analyzing the structure of the data center and cabling used in it. The structure of the data center is based on the TIA-942 and consists of an entrance room, main distribution area MDA, horizontal distribution area HDA , zone distribution area ZDA and the equipment distribution area EDA. Divided into backbone cabling and horizontal cabling and use standard and pre-terminated cabling. The measurement results of structured metal cabling with noise are discussed in the next part. There were used these types of cables UTP, FTP and S/FTP. The measurement was performed in the electricity substation and in the laboratory of electromagnetic compatibility. The third part of this work deals with the actual design of structured cabling for the data center. In the first design we use the standard installation, the second preterminated cabling. For both types of installation cost statement is processed (CAPEX). Thesis is also completed with DC design drawings and rack drawings.

Page generated in 0.0986 seconds