• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 42
  • 17
  • 17
  • 5
  • 5
  • 4
  • 4
  • 3
  • 2
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 118
  • 118
  • 33
  • 23
  • 21
  • 21
  • 18
  • 18
  • 16
  • 16
  • 15
  • 15
  • 13
  • 12
  • 12
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
101

建構物件與物件關聯資料庫工作量產生器 / Toward a Generalized Object-Oriented and Object-Relational Database Workload Generator

黃敏男, Huang, Min-Nan Unknown Date (has links)
本研究以使用者需求導向為設計原則,提出了建構物件與物件關聯資料庫工作量產生器的發展方式,以解決(1)一般績效評估標準規格因只針對特定應用領域,而無法周詳考慮到使用者的特定需求問題;(2)廠商所發展的工作量產生器則因為可能針對其產品的特定領域作設計與測試,而失去了公正性之問題。我們的目地就是希望能以此資料庫工作量產生器的設計方式使測試資料庫效能的工作量,對使用者而言更具一致性、代表性與正確性,如此測試之結果對使用者而言才能更具意義。 我們首先分析SQL3與ODMG標準,以此構成物件關聯與物件導向資料庫之一般化的工作量模式。然後,再著手以使用者需求導向為原則,設計工作量產生器的運作方式。最後,我們實際發展了工作量產生器的雛形系統,以讓使用者能真正依其需求建立工作量並實際作績效測試,並提供給欲建構物件與物件關聯資料庫工作量產生器的人員作參考。 / This thesis provides a method of establishing Object-Oriented and Object-Relational Database Workload Generator, and this method is designed from the perspective of the user's requirements for addressing two problems: (1) the benchmark standards are generally designed to specific domains, so they are unable think comprehensively users' special requirements. (2) the database workload generators that developed by database providers may lose justice, because they might be designed to their products' special domain. Our purpose is that hope this method can make the Workload more consistent, more representative and more accurate and the test result will be more meaningful to users. Firstly, we analyze the SQL3 and ODMG standards to compose the ORDBs' and OODBs' generalized Workload Models. Then, we design our Workload Generator form the perspective of the user's requirements. Finally, we develop the Workload Generator's prototype system, it can allow users to create Workload according to their actual requirements and really do the benchmark test, and furthermore can provide a reference resource to the people who want to develop ORDB and OODB Workload Generator.
102

分散式關聯資料庫系統績效評估工作量模式之研究 / Distributed RDBMS Benchmark Workload Modeling

韓先良, Han, Sien-Liang Unknown Date (has links)
本研究之主要目標在於建構一個能評估分散式關聯資料庫中之特色的需求導向績效評估方法。在過去的績效評估研究中,已經有許多人對於關聯式資料庫績效評估做了多方面的努力。但是,過去的關聯式資料庫資效評估方法如:Wisconsin、AS3AP、TPC系列的Benchmarks都有著一些限制及不足的地方。 過去的關聯式資料庫績效評估方法並無法完全的評估出分散式資料庫的特殊需求及其表現。所以本研究嘗試要建立出一個能專門適用於分散式資料庫導向的績效評估方法。為了要作出此績效評估方法,本研究採用了工作量模式的研究方法。先建出分散式資料庫績效評估的工作量模式,再以其來實作出績效評估方法。工作量模式分成三部分:資料模式、交易模式、控制模式。 / This thesis is intended to design a requirements-centric database benchmark, which can evaluate the general performance of the distributed relational database systems. In the past, there are many relational database benchmarks. But the relational database benchmarks like Wisconsin, AS3AP, TPC, TP1 have some constraints. In this study, we aim to design a general-purpose distributed database workload model and implement it. To design this benchmark, we need to build our workload model. The workload model consists of three components:data model, transaction model, control model. Each model has the requirement specification language to accommodate user's workloads.
103

網際網路資料庫系統績效評估方法之研究--以電子商務為例 / Web database systems benchmark method - electronic commerce orientation

程文成, Cheng, Wen-Cheng Unknown Date (has links)
網路開放商業化以來,各式各樣的商機在網際網路上浮現,大小商家紛在網路上設立據點,或從事廣告、宣傳公司形象,或展示產品內容與產品功能,有的更進一步的設計能與消費者在網路上直接交易,我們概括性的稱之為電子商務。在電子商務的背後,各網站必須要有資料庫作支撐,才能有好的效率和服務。本研究以網際網路資料庫為探討主體,績效為探討目的。並主要以績效評估(Benchmark)理論為貫穿整個研究的基礎理論,一步步地建立適合網際網路資料庫及運算作績效評估的模型。並於模型完成後,予以程式化實作,並以所得數據作測試。 / A lot of business opportunity springs after commercial activities are officially allowed on Internet. The power and the potential of this market is represented by lots of web sites being set up. "Electronic Commerce" is coming to the whole world. However, it is the database supporting business web sites running that counts. Based upon the theory of benchmarking, we want to find out ways that we can know the performance of web database. After the model is set up and the workload is ready, an implementation is being made.
104

Flexibility in Data Management

Voigt, Hannes 07 March 2014 (has links) (PDF)
With the ongoing expansion of information technology, new fields of application requiring data management emerge virtually every day. In our knowledge culture increasing amounts of data and work force organized in more creativity-oriented ways also radically change traditional fields of application and question established assumptions about data management. For instance, investigative analytics and agile software development move towards a very agile and flexible handling of data. As the primary facilitators of data management, database systems have to reflect and support these developments. However, traditional database management technology, in particular relational database systems, is built on assumptions of relatively stable application domains. The need to model all data up front in a prescriptive database schema earned relational database management systems the reputation among developers of being inflexible, dated, and cumbersome to work with. Nevertheless, relational systems still dominate the database market. They are a proven, standardized, and interoperable technology, well-known in IT departments with a work force of experienced and trained developers and administrators. This thesis aims at resolving the growing contradiction between the popularity and omnipresence of relational systems in companies and their increasingly bad reputation among developers. It adapts relational database technology towards more agility and flexibility. We envision a descriptive schema-comes-second relational database system, which is entity-oriented instead of schema-oriented; descriptive rather than prescriptive. The thesis provides four main contributions: (1)~a flexible relational data model, which frees relational data management from having a prescriptive schema; (2)~autonomous physical entity domains, which partition self-descriptive data according to their schema properties for better query performance; (3)~a freely adjustable storage engine, which allows adapting the physical data layout used to properties of the data and of the workload; and (4)~a self-managed indexing infrastructure, which autonomously collects and adapts index information under the presence of dynamic workloads and evolving schemas. The flexible relational data model is the thesis\' central contribution. It describes the functional appearance of the descriptive schema-comes-second relational database system. The other three contributions improve components in the architecture of database management systems to increase the query performance and the manageability of descriptive schema-comes-second relational database systems. We are confident that these four contributions can help paving the way to a more flexible future for relational database management technology.
105

Intelligent Data Layer: : An approach to generating data layer from normalized database model.

Buzo, Amir January 2012 (has links)
Model View Controller (MVC) software architecture is widely spread and commonly used in application’s development. Therefore generation of data layer for the database model is able to reduce cost and time. After research on current Object Relational Mapping (ORM) tools, it was discovered that there are generating tools like Data Access Object (DAO) and Hibernate, however their usage causes problems like inefficiency and slow performance due to many connections with database and set up time. Most of these tools are trying to solve specific problems rather than generating a data layer which is an important component and the bottom layer of database centred applications. The proposed solution to the problem is an engineering approach where we have designed a tool named Generated Intelligent Data Layer (GIDL). GIDL tool generates small models which create the main data layer of the system according to the Database Model. The goal of this tool is to enable and allow software developers to work only with object without deep knowledge in SQL. The problem of transaction and commit is solved by the tool. Also filter objects are constructed for filtering the database. GIDL tool reduced the number of connections and also have a cache where to store object lists and modify them. The tool is compared under the same environment with Hibernate and showed a better performance in terms of time evaluations for the same functions. GIDL tool is beneficial for software developers, because it generates the entire data layer.
106

Performance investigation into selected object persistence stores

Van Zyl, Pieter 21 July 2010 (has links)
The current popular, distributed, n-tiered, object-oriented application architecture pro- vokes many design debates. Designs of such applications are often divided into logical layer (or tiers) - usually user interface, business logic and domain object (or data) layer, each with their own design issues. In particular, the latter contains data that needs to be stored and retrieved from permanent storage. Decisions need to be made as to the most appropriate way of doing this - the choices are usually whether to use an object database, to communicate directly with a relational database, or to use object-relational mapping (ORM) tools to allow objects to be translated to and from their relational form. Most often, depending on the perceived profile of the application, software architects make these decisions using rules of thumb derived from particular experience or the design patterns literature. Although helpful, these rules are often highly context-dependent and are of- ten misapplied. Research into the nature and magnitude of 'design forces' in this area has resulted in a series of benchmarks, intended to allow architects to understand more clearly the implications of design decisions concerning persistence. This study provides some results to help guide the architect's decisions. The study investigated and focused on the <i.performance of object persistence and com- pared ORM tools to object databases. ORM tools provide an extra layer between the business logic layer and the data layer. This study began with the hypothesis that this extra layer and mapping that happens at that point, slows down the performance of object persistence. The aim was to investigate the influence of this extra layer against the use of object databases that remove the need for this extra mapping layer. The study also investigated the impact of certain optimisation techniques on performance. A benchmark was used to compare ORM tools to object databases. The benchmark provided criteria that were used to compare them with each other. The particular benchmark chosen for this study was OO7, widely used to comprehensively test object persistence performance. Part of the study was to investigate the OO7 benchmark in greater detail to get a clearer understanding of the OO7 benchmark code and inside workings thereof. Included in this study was a comparison of the performance of an open source object database, db4o, against a proprietary object database, Versant. These representatives of object databases were compared against one another as well as against Hibernate, a popular open source representative of the ORM stable. It is important to note that these applications were initially used in their default modes (out of the box). Later some optimisation techniques were incorporated into the study, based on feedback obtained from the application developers. There is a common perception that an extra layer as introduced by Hibernate nega- tively impacts on performance. This study showed that such a layer has minimal impact on the performance. With the use of caching and other optimisation techniques, Hibernate compared well against object databases. Versant, a proprietary object database, was faster than Hibernate and the db4o open source object database. Copyright / Dissertation (MSc)--University of Pretoria, 2010. / Computer Science / unrestricted
107

Rozhraní pro aspektové vyhledávání v indexu Wikipedie / Interfaces for Faceted Search in Indexed Wikipedia

Cilip, Peter January 2018 (has links)
Main aim of this thesis is to study existing systems of faceted search and to design own system based on faceted search in the index of Wikipedia. In this thesis we can meet with existing solutions of faceted search. From mistakes and failures of existing solutions was designed our own system, that is output of this thesis. Designed system is described in way of design and implementation. Product of thesis is application and graphical interface. Application interface can be integrated into existing informational system, where it can be used as multidimensional filter. Graphical interface provides option how can application interface be used in real system. System was created focusing on usefullness and simplicity, for using in existing information systems.
108

Flexibility in Data Management

Voigt, Hannes 03 March 2014 (has links)
With the ongoing expansion of information technology, new fields of application requiring data management emerge virtually every day. In our knowledge culture increasing amounts of data and work force organized in more creativity-oriented ways also radically change traditional fields of application and question established assumptions about data management. For instance, investigative analytics and agile software development move towards a very agile and flexible handling of data. As the primary facilitators of data management, database systems have to reflect and support these developments. However, traditional database management technology, in particular relational database systems, is built on assumptions of relatively stable application domains. The need to model all data up front in a prescriptive database schema earned relational database management systems the reputation among developers of being inflexible, dated, and cumbersome to work with. Nevertheless, relational systems still dominate the database market. They are a proven, standardized, and interoperable technology, well-known in IT departments with a work force of experienced and trained developers and administrators. This thesis aims at resolving the growing contradiction between the popularity and omnipresence of relational systems in companies and their increasingly bad reputation among developers. It adapts relational database technology towards more agility and flexibility. We envision a descriptive schema-comes-second relational database system, which is entity-oriented instead of schema-oriented; descriptive rather than prescriptive. The thesis provides four main contributions: (1)~a flexible relational data model, which frees relational data management from having a prescriptive schema; (2)~autonomous physical entity domains, which partition self-descriptive data according to their schema properties for better query performance; (3)~a freely adjustable storage engine, which allows adapting the physical data layout used to properties of the data and of the workload; and (4)~a self-managed indexing infrastructure, which autonomously collects and adapts index information under the presence of dynamic workloads and evolving schemas. The flexible relational data model is the thesis\' central contribution. It describes the functional appearance of the descriptive schema-comes-second relational database system. The other three contributions improve components in the architecture of database management systems to increase the query performance and the manageability of descriptive schema-comes-second relational database systems. We are confident that these four contributions can help paving the way to a more flexible future for relational database management technology.
109

Návrh databázově neutrální objektově-relační vrstvy / Design of a Database Neutral OR Mapper in C++

Ježa, Pavel January 2007 (has links)
This diploma work deals with design and implementation of the database neutral object-relational (OR) layer in C++ language over inherited database. The goal is to create the layer to encase the access to database from the application layer. Suggested layer will stem from the object-relation mapping technology, which is currently available for many object-programming language, such as C#, Java or Visual Basic. The work consists of three main parts. The forepart is focused on clearing object-relation mapping technology. It briefly overviews differences in capabilities and levels of implementation of various approaches. The next part describes significant properties of databases considered as back-ends for data storage in the project. The aim of this part is to present enough information to support database neutral design of the OR layer. The rest of the document deals with design and implementation of OR layer for the considered environment followed by the summarization of results and overall evaluational.
110

Populating a Database to be used with an Indoor Positioning System / Populera en databas som ska användas med ett inomhuspositioneringssystem

Halvarsson, Maria, Qin, Jinglan January 2022 (has links)
Indoor Positioning System (IPS) are becoming more common in many areas such as retail, warehouses, smart facilities, and manufacturing.In recent years, Bluetooth Low Energy (BLE)-based IPS has become increasingly popular due to its low cost and low energy consumption. One of the more recent updates, Bluetooth 5.1, provides the ability to compute the location using Angle of Arrival (AoA) or Angle of Departure (AoD). These new features have allowed for better positioning accuracies, where AoA-based positioning has shown sub-meter accuracy. An application area for BLE-based IPS is retail stores where the technology can benefit both the store and its customers. This thesis investigates how to populate a database to be used with an IPS in a real-life store. The assumption is that customers will have BLE equipped devices and run an application that will send the properly formatted BLE advertisements, such that an BLE IPS can locate the user in the store. Additionally, we assume that the application can use the device's e-compass or other means to determine in which direction the user's device is oriented. Based on the position and orientation of the user, the software is assumed to access a database to know what item(s) are near the customer. However, the question remains of how did this data get into the database? This degree project explores this in detail and assesses the amount of time and effort needed to populate this database and the amount of time and effort needed to keep this database up to date. This project followed an iterative Design Science Research (DSR) methodology where the artifact is the database. A relational database was used as they are widely used and joins can easily be performed and it is easy to modify existing tables. The application was developed in Spring Boot and React. Amazon Web Services (AWS) was used to host and provide the necessary services for the database and application. The result showed that the estimated time needed to populate the database in a supermarket with a sales area of 5300 m2, 36623 products, and 220 containers is 106.64 hours and 107.13 hours in the worst case assuming a walking speed of 1.4 m s-1. Updating a product would take 10.34 s and 10.37 s if the time it takes for a staff member to walk to the place where the product is located is excluded. / Inomhuspositioneringssystem (IPS) har blivit allt vanligare inom många områden. De senaste åren har Bluetooth Low Energy (BLE)-baserad inomhuspositionering blivit mer populärt på grund av egenskaper som låg kostnad och låg energikonsumption. En av de senaste versionerna; Bluetooth 5.1 kan beräkna positionen med hjälp av ankomstvinkeln (eng. Angle of Arrival (AoA)) och utgångsvinkeln (eng. Angle of Departure (AoD)). Angle of Arrival (AoA)-baserad positionering kan uppnå en noggrannhet under 1 m. Ett Bluetooth Low Energy (BLE)-baserat IPS kan tillämpas i butiker där det kan gynna både butikens medarbetare och dess kunder. Detta examensarbete kommer att undersöka hur en databas som ska användas med ett IPS i en butik kan populeras. Antagandet i detta projekt är att kunderna kommer ha BLE utrustade enheter som kör ett program som tillåter att användaren kan lokalisera sig i butiken. Dessutom antar vi att applikationen kan använda enhetens e-kompass eller annan hårdvara för att kunna avgöra i vilken riktning användarens enhet är orienterad. Baserat på användarens position och riktning, antas det att mjukvaran kan komma åt databasen och veta vilka produkter som ligger i närheten av kunden. Dock kvarstår frågan: Hur sparades denna data i databasen? Det här examensarbetet kommer alltså att undersöka hur en databas i en butik kan vara uppbyggd om vi utgår från antagandet att BLE är tillämpat i butiken. Vi kommer även bedöma hur mycket tid som krävs för att populera en databas och hålla denna databas uppdaterad. Ett iterativ arbetsätt tillämpades i examensarbetet. I projektet användes en relationsdatabas eftersom det är enkelt att utföra join-operationer och uppdatera tabbeller i databasen. Applikationen som skapades tillhandahåller funktioner för att spara, uppdatera, radera och söka produkter i databasen. Frontend var kodad i React och backend var kodad i Spring Boot. Amazon Web Services (AWS) molntjänser och deras verktyg användes för att köra applikationen och databasen. Resultatet visade att populera en databas med produkter i en mataffär med en säljyta på 5300 m2, 36623 produkter, och 220 behållare (hyllor, kylskåp, etc.) tog 106.64 timmar i bästa fall och 107.13 timmar i värsta fall om vi antar en gångfart i 1.4 m s-1.

Page generated in 0.1353 seconds