Spelling suggestions: "subject:"association rules"" "subject:"association jules""
1 |
An algorithm for discovering periodical association rulesJiang, Jung-Yi 08 September 2004 (has links)
There are two main contributions in the thesis . Firstly, we design a novel and efficient algorithm for mining calendar-based association rules which have multilevel time granularities in temporal databases. Unlike apriori-like approaches , our method scans the database twice at most. By avoiding multiple scans over the database , our method can reduce the database scanning time.
Secondly, we use membership functions to construct fuzzy calendar patterns which represent asynchronous periods. With the use of fuzzy calendar patterns, we can discover fuzzy periodical association rules which are association rules occurring in asynchronous periods.
Experimental results have shown that our method is more efficient than others, and we can find fuzzy periodical association rules satisfactorily.
|
2 |
arules - A Computational Environment for Mining Association Rules and Frequent Item SetsHornik, Kurt, Grün, Bettina, Hahsler, Michael January 2005 (has links) (PDF)
Mining frequent itemsets and association rules is a popular and well researched approach
for discovering interesting relationships between variables in large databases. The
R package arules presented in this paper provides a basic infrastructure for creating and
manipulating input data sets and for analyzing the resulting itemsets and rules. The package
also includes interfaces to two fast mining algorithms, the popular C implementations
of Apriori and Eclat by Christian Borgelt. These algorithms can be used to mine frequent
itemsets, maximal frequent itemsets, closed frequent itemsets and association rules. (authors' abstract)
|
3 |
Hypothesis-Driven Specialization-based Analysis of Gene Expression Association RulesThakkar, Dharmesh 08 May 2007 (has links)
During the development of many diseases such as cancer and diabetes, the pattern of gene expression within certain cells changes. A vital part of understanding these diseases will come from understanding the factors governing gene expression. This thesis work focused on mining association rules in the context of gene expression. We designed and developed a tool that enables domain experts to interactively analyze association rules that describe relationships in genetic data. Association rules in their native form deal with sets of items and associations among them. But domain experts hypothesize that additional factors like relative ordering and spacing of these items are important aspects governing gene expression. We proposed hypothesis-based specializations of association rules to identify biologically significant relationships. Our approach also alleviates the limitations inherent in the conventional association rule mining that uses a support-confidence framework by providing filtering and reordering of association rules according to other measures of interestingness in addition to support and confidence. Our tool supports visualization of genetic data in the context of a rule, which facilitates rule analysis and rule specialization. The improvement in different measures of interestingness (e.g., confidence, lift, and p-value) enabled by our approach is used to evaluate the significance of the specialized rules.
|
4 |
A computational environment for mining association rules and frequent item setsHahsler, Michael, Grün, Bettina, Hornik, Kurt January 2005 (has links) (PDF)
Mining frequent itemsets and association rules is a popular and well researched approach to discovering interesting relationships between variables in large databases. The R package arules presented in this paper provides a basic infrastructure for creating and manipulating input data sets and for analyzing the resulting itemsets and rules. The package also includes interfaces to two fast mining algorithms, the popular C implementations of Apriori and Eclat by Christian Borgelt. These algorithms can be used to mine frequent itemsets, maximal frequent itemsets, closed frequent itemsets and association rules. (author's abstract) / Series: Research Report Series / Department of Statistics and Mathematics
|
5 |
Universal Design Rules from Product Pairs and Association Rule Based LearningCowen, Nicholas L. 2010 May 1900 (has links)
A product pair is two products with similar functionality that satisfy the same
high level need but are different by design. The goal of this research is to apply
association rule-based learning to product pairs and develop universal design rules to be
used during the conceptual design phase. The Apriori algorithm produced 1,023
association rules with input parameters of 70% minimum confidence and 0.5%
minimum support levels. These rules were down-selected based on the prescribed rule
format of: (Function, Typical User Activity) ? (Change, Universal User Activity). In
other words, for a given product function and user activity, the rules suggest a design
change and new user activity for a more universal product.
This research presents 29 universal design rules to be used during the conceptual
design stage. These universal design rules suggest a parametric, morphological,
functional, or no design change is needed for a given user activity and product function.
No design change rules confirm our intuition and also prevent inefficient design efforts.
A parametric design change is suggested for actionfunction elements involving find hand
use to manipulate a product. Morphological design changes are proposed to solve actionfunction elements in a slightly more complex manner without adding or
subtracting overall functionality. For example, converting human energy to mechanical
energy with the upper body opposed to the lower body or actuating fluid flow with
motion sensors instead of manual knobs. The majority of the recommended functional
changes involve automating a product to make it more universal which might not be
apparently obvious to designers during conceptual design.
|
6 |
Using nuclear receptor interactions as biomarkers for metabolic syndromeHettne, Kristina January 2003 (has links)
<p>Metabolic syndrome is taking epidemic proportions, especially in developed countries. Each risk factor component of the syndrome independently increases the risk of developing coronary artery disease. The risk factors are obesity, dyslipidemia, hypertension, diabetes type 2, insulin resistance, and microalbuminuria. Nuclear receptors is a family of receptors that has recently received a lot of attention due to their possible involvement in metabolic syndrome. Putting the receptors into context with their co-factors and ligands may reveal therapeutic targets not found by studying the receptors alone. Therefore, in this thesis, interactions between genes in nuclear receptor pathways were analysed with the goal of investigating if these interactions can supply leads to biomarkers for metabolic syndrome. Metabolic syndrome donor gene expression data from the BioExpressä, database was analysed with the APRIORI algorithm (Agrawal et al. 1993) for generating and mining association rules. No association rules were found to function as biomarkers for metabolic syndrome, but the resulting rules show that the data mining technique successfully found associations between genes in signaling pathways.</p>
|
7 |
Efficient database management based on complex association rulesZhang, Heng January 2017 (has links)
The large amount of data accumulated by applications is stored in a database. Because of the large amount, name conflicts or missing values sometimes occur. This prevents certain types of analysis. In this work, we solve the name conflict problem by comparing the similarity of the data, and changing the test data into the form of a given template dataset. Studies on data use many methods to discover knowledge from a given dataset. One popular method is association rules mining, which can find associations between items. This study unifies the incomplete data based on association rules. However, most rules based on traditional association rules mining are item-to-item rules, which is a less than perfect solution to the problem. The data recovery system is based on complex association rules able to find two more types of association rules, prefix pattern-to-item, and suffix pattern-to-item rules. Using complex association rules, several missing values are filled in. In order to find the frequent prefixes and frequent suffixes, this system used FP-tree to reduce the time, cost and redundancy. The segment phrases method can also be used for this system, which is a method based on the viscosity of two words to split a sentence into several phrases. Additionally, methods like data compression and hash map were used to speed up the search.
|
8 |
A multilayer framework for quality of context in context-aware systemsAl-Shargabi, Asma Abdulghani Qassem January 2015 (has links)
Context-aware systems use context information to decide what adaptation actions to perform in response to changes in their environment. Depending on applications, context information includes physical context (e.g. temperature and location), user context (e.g. user preferences and user activity), and ICT context (e.g. device capabilities and battery power). Sensors are the main mean of capturing context. Unfortunately, sensed context data are commonly prone to imperfection due to the technical limitations of sensors, their availability, dysfunction, and the highly dynamic nature of environment. Consequently, sensed context data might be imprecise, erroneous, conflicting, or simply missing. To limit the impact of context imperfection on the behavior of a context-aware system, a notion of Quality of Context (QoC) is used to measure quality of any information that is used as context information. Adaptation is performed only if the context data used in the decision-making has an appropriate quality level. This thesis conducts a novel framework for QoC in context-aware systems, which is called MCFQoC (Multilayered-Context Framework for Quality of Context). The main innovative features of our framework, MCFQoC, include: (1) a new definition that generalizes the notion of QoC to encompass sensed context as well as user profiled context; (2) a novel multilayer context model, that distinguishes between three context abstractions: context situation, context object, and context element in descending order. A context element represents a single value and many context elements can be compound into a context object. Many context objects in turn form a context situation; (3) a novel model of QoC parameters which extends the existing parameters with new quality parameter and explicitly distributes the quality parameters across the three layers of context abstraction; (4) a novel algorithm, RCCAR (Resolving Context Conflicts Using Association Rules), which has been developed to resolve conflicts in context data using the Association Rules (AR) technique; (5) a novel mechanism to define QoC policy by assigning weights to QoC parameters using a multi-criteria decision-making technique called Analytical Hierarchy Process (AHP); (6) and finally, a novel quality control algorithm called IPQP (Integrating Prediction with Quality of context Parameters for Context Quality Control) for handling context conflicts, context missing values, and context erroneous values. IPQP is extension of RCCAR. Our framework, MCFQoC, has been implemented in MatLab and evaluated using a case study of a flood forecast system. Results show that the framework is expressive and modular, thanks to the multilayer context model and also to the notion QoC policy which enables us to assign weights for QoC’s parameters depending on quality requirements of each specific application. This flexibility makes it easy to apply our approach to a wider type of context-aware applications. As a part of MCFQoC framework, IPQP algorithm has been successfully tested and evaluated for QoC control using a variety of scenarios. The algorithm RCCAR has been tested and evaluated either individually and as a part of MCFQoC framework with a significant performance concerning resolving context conflicts. In addition, RCCAR has achieved a good success comparing to traditional prediction methods such as moving average (MA), weighted moving average, exponential smoothing, doubled exponential smoothing, and autoregressive moving average (ARMA).
|
9 |
New Probabilistic Interest Measures for Association RulesHahsler, Michael, Hornik, Kurt January 2006 (has links) (PDF)
Mining association rules is an important technique for discovering meaningful patterns in transaction databases. Many different measures of interestingness have been proposed for association rules. However, these measures fail to take the probabilistic properties of the mined data into account. In this paper, we start with presenting a simple probabilistic framework for transaction data which can be used to simulate transaction data when no associations are present. We use such data and a real-world database from a grocery outlet to explore the behavior of confidence and lift, two popular interest measures used for rule mining. The results show that confidence is systematically influenced by the frequency of the items in the left hand side of rules and that lift performs poorly to filter random noise in transaction data. Based on the probabilistic framework we develop two new interest measures, hyper-lift and hyper-confidence, which can be used to filter or order mined association rules. The new measures show significant better performance than lift for applications where spurious rules are problematic. / Series: Research Report Series / Department of Statistics and Mathematics
|
10 |
Implications of probabilistic data modeling for rule miningHahsler, Michael, Hornik, Kurt, Reutterer, Thomas January 2005 (has links) (PDF)
Mining association rules is an important technique for discovering meaningful patterns in transaction databases. In the current literature, the properties of algorithms to mine associations are discussed in great detail. In this paper we investigate properties of transaction data sets from a probabilistic point of view. We present a simple probabilistic framework for transaction data and its implementation using the R statistical computing environment. The framework can be used to simulate transaction data when no associations are present. We use such data to explore the ability to filter noise of confidence and lift, two popular interest measures used for rule mining. Based on the framework we develop the measure hyperlift and we compare this new measure to lift using simulated data and a real-world grocery database. / Series: Research Report Series / Department of Statistics and Mathematics
|
Page generated in 0.0823 seconds