• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 73
  • 22
  • 4
  • 4
  • 3
  • 3
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • 1
  • Tagged with
  • 136
  • 136
  • 44
  • 41
  • 33
  • 24
  • 23
  • 22
  • 19
  • 19
  • 19
  • 18
  • 17
  • 15
  • 15
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
101

Data Mining For Rule Discovery In Relational Databases

Toprak, Serkan 01 September 2004 (has links) (PDF)
Data is mostly stored in relational databases today. However, most data mining algorithms are not capable of working on data stored in relational databases directly. Instead they require a preprocessing step for transforming relational data into algorithm specified form. Moreover, several data mining algorithms provide solutions for single relations only. Therefore, valuable hidden knowledge involving multiple relations remains undiscovered. In this thesis, an implementation is developed for discovering multi-relational association rules in relational databases. The implementation is based on a framework providing a representation of patterns in relational databases, refinement methods of patterns, and primitives for obtaining necessary record counts from database to calculate measures for patterns. The framework exploits meta-data of relational databases for pruning search space of patterns. The implementation extends the framework by employing Apriori algorithm for further pruning the search space and discovering relational recursive patterns. Apriori algorithm is used for finding large itemsets of tables, which are used to refine patterns. Apriori algorithm is modified by changing support calculation method for itemsets. A method for determining recursive relations is described and a solution is provided for handling recursive patterns using aliases. Additionally, continuous attributes of tables are discretized utilizing equal-depth partitioning. The implementation is tested with gene localization prediction task of KDD Cup 2001 and results are compared to those of the winner approach.
102

Design and implementation of a database programming language for XML-based applications

Schuhart, Henrike January 2006 (has links)
Zugl.: Lübeck, Univ., Diss., 2006
103

Contributions à la mise en place d'une infrastructure de Cloud Computing à large échelle / Contributions to massively distributed Cloud Computing infrastructures

Pastor, Jonathan 18 October 2016 (has links)
La croissance continue des besoins en puissance de calcul a conduit au triomphe du modèle de Cloud Computing. Des clients demandeurs en puissance de calcul vont s’approvisionner auprès de fournisseurs d’infrastructures de Cloud Computing, mises à disposition via Internet. Pour réaliser des économies d’échelles, ces infrastructures sont toujours plus grandes et concentrées en quelques endroits, conduisant à des problèmes tels que l’approvisionnement en énergie, la tolérance aux pannes et l’éloignement des utilisateurs. Cette thèse s’est intéressée à la mise en place d’un système d’IaaS massivement distribué et décentralisé exploitant un réseau de micros centres de données déployés sur la dorsale Internet, utilisant une version d’OpenStack revisitée pendant cette thèse autour du support non intrusif de bases de données non relationnelles. Des expériences sur Grid’5000 ont montré des résultats intéressants sur le plan des performances, toutefois limités par le fait qu’OpenStack ne tirait pas avantage nativement d’un fonctionnement géographiquement réparti. Nous avons étudié la prise en compte de la localité réseau pour améliorer les performances des services distribués en favorisant les collaborations proches. Un prototype de l’algorithme de placement de machines virtuelles DVMS, fonctionnant sur une topologie non structurée basée sur l’algorithme Vivaldi, a été validé sur Grid’5000. Ce prototype a fait l’objet d’un prix scientifique lors de l’école de printemps Grid’50002014. Enfin, ces travaux nous ont amenés à participer au développement du simulateur VMPlaceS. / The continuous increase of computing power needs has favored the triumph of the Cloud Computing model. Customers asking for computing power will receive supplies via Internet resources hosted by providers of Cloud Computing infrastructures. To make economies of scale, Cloud Computing that are increasingly large and concentrated in few attractive places, leading to problems such energy supply, fault tolerance and the fact that these infrastructures are far from most of their end users. During this thesis we studied the implementation of an fully distributed and decentralized IaaS system operating a network of micros data-centers deployed in the Internet backbone, using a modified version of OpenStack that leverages non relational databases. A prototype has been experimentally validated onGrid’5000, showing interesting results, however limited by the fact that OpenStack doesn’t take advantage of a geographically distributed functioning. Thus, we focused on adding the support of network locality to improve performance of Cloud Computing services by favoring collaborations between close nodes. A prototype of the DVMS algorithm, working with an unstructured topology based on the Vivaldi algorithm, has been validated on Grid’5000. This prototype got the first prize at the large scale challenge of the Grid’5000 spring school in 2014. Finally, the work made with DVMS enabled us to participate at the development of the VMPlaceS simulator.
104

Porovnání schématu relační databáze a struktur formátu XML / Comparison of relational database schema and XML structures

Vodňanský, Daniel January 2013 (has links)
The work deals with the relationship of the relational model and XML schema document and its technological and pragmatic aspects. It defines the theoretical field of data modeling at conceptual level and the two mentioned possible implementation models at the physical level. The aim is to answer the question when in the design and development of application or system it is appropriate to proceed with one of these models. Furthermore, this work also provides a general procedure for mapping conceptual schema into XML schema structures and solutions to problems that can come across during the mapping process. The problem is solved by analyzing two real issues - timetables of public transportation and the information system of a swimming school, formalized through a mechanism of predicate logic. Unlike most works on a similar topic this one varies in a pragmatic view on the problem - the concept of data, their origin, their target user and structuring.
105

Billing and receivables database application

Lukalapu, Sushma 01 January 2000 (has links)
The purpose of this project is to design, build, and implement an information retrieval database system for the Accounting Department at CSUSB. The database will focus on the financial details of the student accounts maintained by the accounting personnel. It offers detailed information pertinent to tuition, parking, housing, boarding, etc.
106

Anomaly Detection Techniques for the Protection of Database Systems against Insider Threats

Asmaa Mohamed Sallam (6387488) 15 May 2019 (has links)
The mitigation of insider threats against databases is a challenging problem since insiders often have legitimate privileges to access sensitive data. Conventional security mechanisms, such as authentication and access control, are thus insufficient for the protection of databases against insider threats; such mechanisms need to be complemented with real-time anomaly detection techniques. Since the malicious activities aiming at stealing data may consist of multiple steps executed across temporal intervals, database anomaly detection is required to track users' actions across time in order to detect correlated actions that collectively indicate the occurrence of anomalies. The existing real-time anomaly detection techniques for databases can detect anomalies in the patterns of referencing the database entities, i.e., tables and columns, but are unable to detect the increase in the sizes of data retrieved by queries; neither can they detect changes in the users' data access frequencies. According to recent security reports, such changes are indicators of potential data misuse and may be the result of malicious intents for stealing or corrupting the data. In this thesis, we present techniques for monitoring database accesses and detecting anomalies that are considered early signs of data misuse by insiders. Our techniques are able to track the data retrieved by queries and sequences of queries, the frequencies of execution of periodic queries and the frequencies of referencing the database tuples and tables. We provide detailed algorithms and data structures that support the implementation of our techniques and the results of the evaluation of their implementation.<br>
107

Upotreba fazi logike u relacionim bazama podataka / Fuzzy logic usage in relational databases

Škrbić Srđan 19 March 2009 (has links)
<p>Doktorska disertacija pripada oblasti<br />informacionih sistema, odnosno podoblasti koja<br />se bavi upravljanjem skladi&scaron;tenjem i<br />pretraživanjem informacija. Osnovni cilj<br />disertacije je modeliranje i implementacija<br />skupa alata koji omogućavaju upotrebu fazi<br />logike u radu sa relacionim bazama podataka.<br />Da bi se do tog skupa alata do&scaron;lo, najpre je<br />relacioni model podataka pro&scaron;iren elementima<br />teorije fazi skupova, a zatim je definisano fazi<br />pro&scaron;irenje upitnog jezika SQL &ndash; PFSQL.<br />Interpreter za taj jezik je implementiran u<br />okviru fazi JDBC drajvera koji, osim<br />implementacije interpretera, sadrži i elemente<br />koji omogućavaju jednostavnu upotrebu ovih<br />mehanizama iz programskog jezika Java. Skup<br />alata je zaokružen implementacijom CASE<br />alata za razvoj fazi-relacionog modela baze<br />podataka. Osim toga, razmatrane su i<br />mogućnosti za upotrebu PFSQL jezika u<br />vi&scaron;eslojnim aplikacijama.</p> / <p>This doctoral dissertation belongs to the<br />field of information systems, subfield<br />information storage and retrieval management.<br />The main subject of the dissertation is modeling<br />and implementation of a set of tools that allow<br />usage of fuzzy logic in relational database<br />applications<br />In order to achieve that goal, at first, the<br />relational data model is extended with elements<br />of fuzzy set theory. After that, a fuzzy<br />extension of the SQL query language, called<br />PFSQL, is defined. An interpreter for that<br />language is implemented as a part of the fuzzy<br />JDBC driver. Beside the implementation of the<br />interpreter, this fuzzy JDBC driver contains<br />elements that allow simple usage of offered<br />mechanisms from Java programming language.<br />The set of tools is concluded with the<br />implementation of the CASE tool for the<br />development of fuzzy-relational data models. In<br />addition, possibilities to use PFSQL language<br />on the middle tier of multi tier systems are<br />discussed.</p>
108

Utilization of ETL Processes for Geographical Data Migration : A Case Study at Metria AB

Sihvola, Toni January 2024 (has links)
In this study, the safety of using ETL processes to migrate geographical data between heterogeneous data sources was investigated, as well as whether certain data structures are more prone to integrity loss during such migrations. Geographical data in various vector structures was migrated using ETL software, FME, from a legacy data source (Oracle 11g with integrated Esri geodatabases) to another (PostgreSQL 14.10 with the PostGIS extension) in order to explore the aforementioned challenges. The maintenance of data integrity post-migration was assessed by comparing the difference between the geodata housed in Oracle 11g (the source) and PostgreSQL 14.10 (the destination) using ArcGIS Pro's built-in tools and a Python script. Further evaluation of the role of ETL processes in geographical data migration included conducting interviews with specialists in databases, data migration, and FME both before and after the migration. The study concludes that different vector structures are affected differently. Whereas points and lines maintained 100% data integrity across all datasets, polygons achieved 99.95% accuracy in one out of the three tested datasets. Managing this issue can be addressed by implementing a repair process during the Transform stage of an ETL process. However, such a process does not guarantee an entirely successful outcome; although the affected area was significantly reduced post-repair, the polygons contained a higher amount of mismatches. / I denna studie undersöktes om ETL-processer kan användas på ett säkert sätt för att migrera geografiska data mellan heterogena datakällor, samt om vissa datastrukturer är mer benägna att förlora integritet under sådana migrationer. Geografiskt data i olika vektorstrukturer migrerades med hjälp av ETL-programvaran FME, från en föråldrad datakälla (Oracle 11g med integrerade Esri geodatabaser) till en annan (PostgreSQL 14.10 med PostGIS-tillägget) för att utforska de ovannämnda frågorna. Dataintegritet mättes genom att jämföra skillnaden mellan geodatan på Oracle 11g (källan) och PostgreSQL 14.10 (destinationen) med hjälp av ArcGIS Pro's inbyggda verktyg och ett Python skript. För att ytterligare utvärdera rollen av ETL-processer i migrering av geografiskt data genomfördes intervjuer med specialister inom databaser, datamigration och FME, både före och efter migrationen.  Studien konstaterar att olika vektorstrukturer påverkas olika. Medan punkter och linjer bibehöll 100% datatillförlitlighet över alla dataset, uppnådde polygoner 99,95% noggrannhet i ett av de tre testade dataseten. Hantering av detta problem kan adresseras genom att implementera en reparationsprocess under Transform-steget av en ETL-process. Dock garanterar inte en sådan process ett helt lyckat resultat; även om den påverkade arean minskades avsevärt efter reparationen, innehöll polygonerna ett högre antal avvikelser.
109

Desenvolvimento de um protótipo de sistema de suporte a decisão baseado em alertas vinculado a um sistema de informações sobre medicamentos cardiovasculares

Neves, Eugenio Rodrigo Zimmer January 1995 (has links)
O incentivo ao uso racional de medicamentos fator importante na melhoria das condições de saúde. A efetivação desta racionalidade, no entanto, esbarra tanto em fatores culturais como em fatores de escassez ou mesmo ausência de informação confiável. Os farmacêuticos, membros indispensáveis de qualquer equipe multidisciplinar de saúde, cumprem papel preponderante na disseminação deste conhecimento especializado. Com o objetivo de suprir as lacunas existentes quanto a qualidade da informação farmacológico-terapêutica existente no Brasil, este trabalho desenvolve um sistema de informações sobre medicamentos cardiovasculares aliado a um Sistema de Suporte a Decisão Baseado em Alertas, utilizando uma arquitetura que combina bases de dados relacionais com Medical Logical Modules. O sistema desenvolvido proporciona, a farmacêuticos e outros profissionais da área da saúde, não apenas consultas a informação, mas também sugestões e alertas contextuais referentes ao uso correto de medicamentos, contribuindo, assim, para o treinamento no próprio trabalho destes. / The rational use of drugs is an important factor to the health conditions improvement. The pharmacists are responsible for advising customers about the rational use of drugs, but this professional effort is limited by cultural factors, absence or lack of reliable infomation. The aim of this work is to support pharmacists through a cardiovascular medicine Information System attached to an Alert-based Decision Support System whose structure combines Medical Logical Modules and relational databases. The developed system not only provides the health professional and pharmacist with drug information, but also suggests or alerts them about the proper use of drugs as well it contributes to their in site training.
110

資料交換與查詢在XML文件與關連資料庫之間 / Data Exchange and Query Language between XML Documents and Relational Databases

王瑞娟 Unknown Date (has links)
隨著全球資訊網(World Wild Web,簡稱WWW或Web)的日趨普及,我們發現愈來愈多的資料是直接從網路上呈現與存取的。不同於過去關聯式資料庫(Relational Database Management Systems,RDBMS)的結構式資料(Structured Data)。現今許多資料都是直接以HTML(Hypertext Markup Language)格式呈現,然而HTML 的標籤只是在做資料的呈現。為了讓網際網路上的資料可真正順利傳達於組織間,新興的XML逐漸受到重視。相較於HTML,XML標籤是在做資料的定義,讓定義好的資料直接透過網際網路傳達於組織間,具有在組織間再使用(Reuse)的能力,因此現在逐漸成為組織間資料整合與轉換時一個好的解決方式。但面對傳統的關連式資料庫又該如何與XML文件整合(Data Integration)的動作,讓兩種不同來源的資料能夠相互運算(Interoperability),達成異質性資料的同質化(Homogeneous)的功效。使不同來源的資料可雙向的互相溝通,是目前急欲被探討的問題。因此本研究便發展了對關聯式資料庫與XML文件兩種來源相互轉換的溝通機制,讓資料能在這兩種來源間相互交換與利用。 / With the popularity of WWW( World Wild Web or Web ),we have seen large volume of data is available on the Web. Different from the data stored in traditional RDBMS (Relational Databases Managements Systems) which is structured data, huge data now is stored directly in the form of HTML (Hypertext Markup Language) pages. For representing data and interchange data between multiple data sources on the Web, XML (Extensible Markup Language) is a fast emerging as the dominant standard. Like HTML, XML is a subset of SGML. However, whereas HTML tags serve the primary purpose of describing how to display a data item, XML tags describe the data itself. The initial impetus for XML may have been primarily to enhance this ability of remote applications to interpret and operate on documents fetched over the Internet, so it has become the best solution now to solve the problems with data exchange and translation between the multiple sources. But XML also raises a problem: how to integrate XML documents with data stored in the traditional RDBMS. The objective is to communicate bi-directional data sources between RDBMS data and XML documents, and has the ability to interoperate data between multiple data sources. Finally, reach the purpose of heterogeneous data become homogeneous. In this research, we try to develop a translation model between RDBMS data and XML documents, in order to exchange and reuse data between different sources.

Page generated in 0.4019 seconds