• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 245
  • 27
  • 19
  • 12
  • 8
  • 8
  • 6
  • 3
  • 3
  • 2
  • 2
  • 2
  • 2
  • 1
  • Tagged with
  • 391
  • 135
  • 79
  • 63
  • 62
  • 57
  • 55
  • 52
  • 49
  • 48
  • 46
  • 40
  • 35
  • 34
  • 34
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
61

Testing scalability of cloud gaming for multiplayer game

Printzell, Dan January 2018 (has links)
Background. The rendering of games takes a lot of processing power and requires expensivehardware to be able to perform this task in a real-time with an acceptableframe-rate. Games often also require an anti-cheat system that require extrapower to be able to always verify that the game has not been modified. With the help ofgame streaming these disadvantages could be removed from the clients. Objectives. The objective of this thesis is to create a game streaming server and client tosee if a game streaming server could scale with the amount of coresit has access to. Methods. The research question will be answered using the implementation methodology, and an experiment will be conducted using that implementation. Two programs are implemented, the server program and the client program.The servers implement the management of clients, the game logic, the rendering and the compression. Each client can only be connected to one server and the server and its clients live inside of a game instance. Everyone that is connected to one server play on the same instance.The implementation is implemented in the D programming language, and it uses the ZLib and the SDL2 libraries as the building blocks.With all of these connected an experiment is designed where as many clients as possible connect to the server. With this data a plot is create in the result section. Results. The output data shows that the implementation scale and a formula was made-up to match the scalability. The formula is <img src="http://www.diva-portal.org/cgi-bin/mimetex.cgi?f(x)%20=%208%20+%205x%20-%200.11x%5E2" />. Conclusions. The experiment was successful and showed that the game server successfully scaledbased on the number of cores that where allocated. It does not scale as good as expected,but it is still an success. The test results are limited as it was only tested on one setup. More research is needed to test it on more hardware and to be able find more optimized implementations.
62

Performance and Scaling Analysis of a Hypocycloid Wiseman Engine

January 2014 (has links)
abstract: The slider-crank mechanism is popularly used in internal combustion engines to convert the reciprocating motion of the piston into a rotary motion. This research discusses an alternate mechanism proposed by the Wiseman Technology Inc. which involves replacing the crankshaft with a hypocycloid gear assembly. The unique hypocycloid gear arrangement allows the piston and the connecting rod to move in a straight line, creating a perfect sinusoidal motion. To analyze the performance advantages of the Wiseman mechanism, engine simulation software was used. The Wiseman engine with the hypocycloid piston motion was modeled in the software and the engine's simulated output results were compared to those with a conventional engine of the same size. The software was also used to analyze the multi-fuel capabilities of the Wiseman engine using a contra piston. The engine's performance was studied while operating on diesel, ethanol and gasoline fuel. Further, a scaling analysis on the future Wiseman engine prototypes was carried out to understand how the performance of the engine is affected by increasing the output power and cylinder displacement. It was found that the existing Wiseman engine produced about 7% less power at peak speeds compared to the slider-crank engine of the same size. It also produced lower torque and was about 6% less fuel efficient than the slider-crank engine. These results were concurrent with the dynamometer tests performed in the past. The 4 stroke diesel variant of the same Wiseman engine performed better than the 2 stroke gasoline version as well as the slider-crank engine in all aspects. The Wiseman engine using contra piston showed poor fuel efficiency while operating on E85 fuel. But it produced higher torque and about 1.4% more power than while running on gasoline. While analyzing the effects of the engine size on the Wiseman prototypes, it was found that the engines performed better in terms of power, torque, fuel efficiency and cylinder BMEP as their displacements increased. The 30 horsepower (HP) prototype, while operating on E85, produced the most optimum results in all aspects and the diesel variant of the same engine proved to be the most fuel efficient. / Dissertation/Thesis / M.S.Tech Mechanical Engineering 2014
63

Resource Allocation Guidelines : Configuring a large telecommunication application / Riktlinjer för resursallokering

Eriksson, Daniel January 2001 (has links)
Changing the architecture of the Ericsson Billing Gateway application has shown to solve the problem with dynamic memory management that decreased the performance. The new architecture that is focused on processes instead of threads showed increased performance. It also allowed for the possibility to adjust the process / thread configuration towards the network topology and hardware. Measurements of different configurations showed the importance of an accurate configuration and also that certain guidelines could be established based on the results. / Genom att ändra architecturen på Billing Gateway löste man på Ericsson problemet med &quot;Dynamisk minneshantering&quot;. Den nya architecturen som fokuserar på processer istället för trådar visade ökad prestanda. Den nya architecturen tillät också vissa konfigurationsmöjligheter gentemot nätverkstopologi och hårdvara. Mätningar på olika konfigurationer visade vikten av rätt konfiguration och också att vissa riktlinjer kunde utrönas från resultaten. / Phone#: +46 457 66582
64

Dependability aspects of COM+ and EJB in multi-tiered distributed systems

Karásson, Robert January 2002 (has links)
COM+ and Enterprise JavaBeans are two component-based technologies that can be used to build enterprise systems. These are two competing technologies in the software industry today and choosing which technology a company should use to build their enterprise system is not an easy task. There are many factors to consider and in this project we evaluate these two technologies with focus on scalability and the dependability aspects security, availability and reliability. Independently, these technologies are theoretically evaluated with the criteria in mind. We use a 4-tier architecture for the evaluation and the center of attention is a persistence layer, which typically resides in an application server, and how it can be realized using the technologies. This evaluation results in a recommendation about which technology is a better approach to build a scalable and dependable distributed system. The results are that COM+ is considered a better approach to build this kind of multi-tier distributed systems.
65

Limitations of Azure in GIS Scalability : A performance and migration study

Bäckström, Jonas January 2012 (has links)
In this study, the cloud platform Windows Azure has been targeted for test implementations of Geographical Information System (GIS) software in the form of map servers and tile caches. The map servers included were GeoServer, MapNik, MapServer and SharpMap, which together with the tile caches, GeoWebCache, MapCache and TileCache, were installed on Windows Azures three different virtual machine roles (Web, Worker and VM). Furthermore, different techniques for scalingapplications and internal role communication are presented, followed by four sets of performance tests. The performance tests attempt to highlight the differences in request times, how the different role sizes handle the load from the incoming requests, how the different role sizes handle many concurrent TCP-connections and how well the incoming requests are load balanced in between the worker roles. The test implementations showed that all map servers and tile caches were successfully installed in Azure, which leads to the conclusion that Windows Azure is suitable for hosting GIS software with similar installation requirements to the previously mentioned software. Four different approaches (Direct mapping, Public Internal Endpoints, Queue and Worker Role Request Broker) are presented showing how Azure allows different methods in order to scale the internal role communication as well as the external client requests. The performance tests provided somewhat inconclusive test results due to hardware limitations in the test setup. This made it difficult to draw concluding parallels between the final results and the expected values. Minor tendencies in performance gain can be seen when scaling the VM size as well as the number of VMs.
66

Adapting video compression to new formats / Adaptation de la compression vidéo aux nouveaux formats

Bordes, Philippe 18 January 2016 (has links)
Les nouvelles techniques de compression vidéo doivent intégrer un haut niveau d'adaptabilité, à la fois en terme de bande passante réseau, de scalabilité des formats (taille d'images, espace de couleur…) et de compatibilité avec l'existant. Dans ce contexte, cette thèse regroupe des études menées en lien avec le standard HEVC. Dans une première partie, plusieurs adaptations qui exploitent les propriétés du signal et qui sont mises en place lors de la création du bit-stream sont explorées. L'étude d'un nouveau partitionnement des images pour mieux s'ajuster aux frontières réelles du mouvement permet des gains significatifs. Ce principe est étendu à la modélisation long-terme du mouvement à l'aide de trajectoires. Nous montrons que l'on peut aussi exploiter la corrélation inter-composantes des images et compenser les variations de luminance inter-images pour augmenter l'efficacité de la compression. Dans une seconde partie, des adaptations réalisées sur des flux vidéo compressés existants et qui s'appuient sur des propriétés de flexibilité intrinsèque de certains bit-streams sont investiguées. En particulier, un nouveau type de codage scalable qui supporte des espaces de couleur différents est proposé. De ces travaux, nous dérivons des metadata et un modèle associé pour opérer un remapping couleur générique des images. Le stream-switching est aussi exploré comme une application particulière du codage scalable. Plusieurs de ces techniques ont été proposées à MPEG. Certaines ont été adoptées dans le standard HEVC et aussi dans la nouvelle norme UHD Blu-ray Disc. Nous avons investigué des méthodes variées pour adapter le codage de la vidéo aux différentes conditions de distribution et aux spécificités de certains contenus. Suivant les scénarios, on peut sélectionner et combiner plusieurs d'entre elles pour répondre au mieux aux besoins des applications. / The new video codecs should be designed with an high level of adaptability in terms of network bandwidth, format scalability (size, color space…) and backward compatibility. This thesis was made in this context and within the scope of the HEVC standard development. In a first part, several Video Coding adaptations that exploit the signal properties and which take place at the bit-stream creation are explored. The study of improved frame partitioning for inter prediction allows better fitting the actual motion frontiers and shows significant gains. This principle is further extended to long-term motion modeling with trajectories. We also show how the cross-component correlation statistics and the luminance change between pictures can be exploited to increase the coding efficiency. In a second part, post-creation stream adaptations relying on intrinsic stream flexibility are investigated. In particular, a new color gamut scalability scheme addressing color space adaptation is proposed. From this work, we derive color remapping metadata and an associated model to provide low complexity and general purpose color remapping feature. We also explore the adaptive resolution coding and how to extend scalable codec to stream-switching applications. Several of the described techniques have been proposed to MPEG. Some of them have been adopted in the HEVC standard and in the UHD Blu-ray Disc. Various techniques for adapting the video compression to the content characteristics and to the distribution use cases have been considered. They can be selected or combined together depending on the applications requirements.
67

Finding Community Structures In Social Activity Data

Peng, Chengbin 19 May 2015 (has links)
Social activity data sets are increasing in number and volume. Finding community structure in such data is valuable in many applications. For example, understand- ing the community structure of social networks may reduce the spread of epidemics or boost advertising revenue; discovering partitions in tra c networks can help to optimize routing and to reduce congestion; finding a group of users with common interests can allow a system to recommend useful items. Among many aspects, qual- ity of inference and e ciency in finding community structures in such data sets are of paramount concern. In this thesis, we propose several approaches to improve com- munity detection in these aspects. The first approach utilizes the concept of K-cores to reduce the size of the problem. The K-core of a graph is the largest subgraph within which each node has at least K connections. We propose a framework that accelerates community detection. It first applies a traditional algorithm that is relatively slow to the K-core, and then uses a fast heuristic to infer community labels for the remaining nodes. The second approach is to scale the algorithm to multi-processor systems. We de- vise a scalable community detection algorithm for large networks based on stochastic block models. It is an alternating iterative algorithm using a maximum likelihood ap- proach. Compared with traditional inference algorithms for stochastic block models, our algorithm can scale to large networks and run on multi-processor systems. The time complexity is linear in the number of edges of the input network. The third approach is to improve the quality. We propose a framework for non- negative matrix factorization that allows the imposition of linear or approximately linear constraints on each factor. An example of the applications is to find community structures in bipartite networks, which is useful in recommender systems. Our algorithms are compared with the results in recent papers and their quality and e ciency are verified by experiments.
68

Scalability Guidelines for Software as a Service: Recommendations based on a case analysis of Quinyx FlexForce AB and UCMS Group Ltd.

Rapp, Mikael January 2010 (has links)
Software as a service (SaaS) has become a common business model for application providers. However, this success has lead to scalability issues for service providers. Their user base and processing needs can grow rapidly and it is not always clear how a SaaS provider should optimally scale their service. This thesis summarizes the technological (both software and hardware) related solutions to scaling, but will also cover financial and managerial aspects of scalability. Unfortunately, there is no existing out of the box solution for managing scalability, every situation and application is currently viewed as a unique problem, but there exists a lot of good advice from many sources about scaling. Obviously there are practical solutions to scaling, as there are successful SaaS providers, but it is not clear if there exists some fundamental principles that every SaaS provider could use to address the issue of scalability. This thesis seeks to find such fundamental principles though previous research, articles and finally a case analysis of Quinyx FlexForce AB. The thesis concludes that there are many principles of scaling a 3-tier web system and that most principles can be applied by SaaS providers. / Software as a Service (SaaS) har blivit en allt vanligare lösning för företag. Detta har dock lett till skalbarhets problem för många leverantörer av SaaS. En SaaS leverantör kan få problem med skalning ifall deras användarbas eller beräkningsbehov växer för snabbt. Denna avhandling sammanfattar de tekniska (mjukvara och hårdvara) relaterade lösningar på skalning. Avhandlingen kommer även kortfattat att omfatta ekonomiska och administrativa aspekter av skalbarhet. Tyvärr finns det inga befintliga universallösningar för hantering skalbarhet utan varje situation och tillämpning måste ses som ett unikt problem. Det finns många goda råd från många källor om skalning och uppenbarligen finns det praktiska lösningar på att skala, då det finns framgångsrika SaaS leverantörer. Det är dock oklart om det finns några grundläggande principer som varje SaaS-leverantör kan använda för att underlätta skalbarhet. Avhandlingen syftar till att hitta sådana grundläggande principer och grundar sig på tidigare forskning, aktuella artiklar och avslutats med en analys av Quinyx FlexForce AB. Avhandlingen drar slutsatsen att det finns grundläggande principer som SaaS leverantörer kan tillämpa vid skalning av ett ”3-tier” webserver system.
69

Scalable Stream Processing and Management for Time Series Data

Mousavi, Bamdad 15 June 2021 (has links)
There has been an enormous growth in the generation of time series data in the past decade. This trend is caused by widespread adoption of IoT technologies, the data generated by monitoring of cloud computing resources, and cyber physical systems. Although time series data have been a topic of discussion in the domain of data management for several decades, this recent growth has brought the topic to the forefront. Many of the time series management systems available today lack the necessary features to successfully manage and process the sheer amount of time series being generated today. In this today we stive to examine the field and study the prior work in time series management. We then propose a large system capable of handling time series management end to end, from generation to consumption by the end user. Our system is composed of open-source data processing frameworks. Our system has the capability to collect time series data, perform stream processing over it, store it for immediate and future processing and create necessary visualizations. We present the implementation of the system and perform experimentations to show its scalability to handle growing pipelines of incoming data from various sources.
70

Evaluation of CockroachDB in a cloud-native environment

Håkansson, Kristina, Rosenqvist, Andreas January 2021 (has links)
The increased demand for using large databases that scale easily and stay consistent requires service providers to find new solutions for storing data in databases. One solution that has emerged is cloud-native databases. Service providers who effectively can transit to cloud-native databases will benefit from new enterprise applications, industrial automation, Internet of Things (IoT) as well as consumer services, such as gaming and AR/VR. This consequently changes the requirements on a database's architecture and infrastructure in terms of being compatible with the services deployed in a cloud-native environment - this is where CockroachDB comes into the picture. CockroachDB is relatively new and is built from the ground up to run in a cloud-native environment. It is built up with nodes that work as individual machines, and these nodes form a cluster. The authors of this report aim to evaluate the characteristics of the Cockroach database to get an understanding of what it offers to companies that are in a cloud-infrastructure transition phase. For the scope of characteristics, this report is focusing on performance, throughput, stress-test, version hot-swapping, horizontal-/vertical scaling, and node disruptions. To do this, a CockroachDB database was deployed on a Kubernetes cluster, in which simulated traffic was conducted. For the throughput measurement, the TPC-C transaction processing benchmark was used. For scaling, version hot-swapping, and node disruptions, an experimental method was performed. The result of the study confirms the expected outcome. CockroachDB does in fact scale easily, both horizontally and vertically, with minimal effort. It also shows that the throughput remains the same when the cluster is scaled up and out since CockroachDB does not have a master write-node, which is the case with some other databases. CockroachDB also has built-in functionality to handle configuration changes like version hot-swapping and node disruptions. This study concluded that CockroachDB lives up to its promises regarding the subjects handled in the report, and can be seen as a robust, easily scalable database that can be deployed in acloud-native environment.

Page generated in 0.0579 seconds