561 |
Kato: a plagiarism-detection tool for answer-set programsOetsch, Johannes, Schwengerer, Martin, Tompits, Hans January 2010 (has links)
We present the tool Kato which is, to the best of our knowledge, the first tool for plagiarism detection that is directly tailored for answer-set programming (ASP). Kato aims at finding similarities between (segments of) logic programs to help detecting cases of plagiarism. Currently, the tool is realised for DLV programs but it is designed to handle various logic-programming syntax versions. We review basic features and the underlying methodology of the tool.
|
562 |
Constraint-based abstraction of a model checker for infinite state systemsBanda, Gourinath, Gallagher, John P. January 2010 (has links)
Abstract interpretation-based model checking provides an approach to verifying properties of infinite-state systems. In practice, most previous work on abstract model checking is either restricted to verifying universal properties, or develops special techniques for temporal logics such as modal transition systems or other dual transition systems. By contrast we apply completely standard techniques for constructing abstract interpretations to the abstraction of a CTL semantic function, without restricting the kind of properties that can be verified. Furthermore we show that this leads directly to implementation of abstract model checking algorithms for abstract domains based on constraints, making use of an SMT solver.
|
563 |
Range restriction for general formulasBrass, Stefan January 2010 (has links)
Deductive databases need general formulas in rule bodies, not only conjuctions of literals. This is well known since the work of Lloyd and Topor about extended logic programming. Of course, formulas must be restricted in such a way that they can be effectively evaluated in finite time, and produce only a finite number of new tuples (in each iteration of the TP-operator: the fixpoint can still be infinite). It is also necessary to respect binding restrictions of built-in predicates: many of these predicates can be executed only when certain arguments are ground. Whereas for standard logic programming rules, questions of safety, allowedness, and range-restriction are relatively easy and well understood, the situation for general formulas is a bit more complicated. We give a syntactic analysis of formulas that guarantees the necessary properties.
|
564 |
Transforming imperative algorithms to constraint handling rulesAbdennadher, Slim, Ismail, Haythem, Khoury, Frederick January 2010 (has links)
Different properties of programs, implemented in Constraint Handling Rules (CHR), have already been investigated. Proving these properties in CHR is fairly simpler than proving them in any type of imperative programming language, which triggered the proposal of a methodology to map imperative programs into equivalent CHR. The equivalence of both programs implies that if a property is satisfied for one, then it is satisfied for the other.
The mapping methodology could be put to other beneficial uses. One such use is the automatic generation of global constraints, at an attempt to demonstrate the benefits of having a rule-based implementation for constraint solvers.
|
565 |
Persistent constraints in constraint handling rulesBetz, Hariolf, Raiser, Frank, Frühwirth, Thom January 2010 (has links)
In the most abstract definition of its operational semantics, the declarative and concurrent programming language CHR is trivially non-terminating for a significant class of programs. Common refinements of this definition, in closing the gap to real-world implementations, compromise on declarativity and/or concurrency. Building on recent work and the notion of persistent constraints, we introduce an operational semantics avoiding trivial non-termination without compromising on its essential features.
|
566 |
A tool for generating partition schedules of multiprocessor systemsGoltz, Hans-Joachim, Pieth, Norbert January 2010 (has links)
A deterministic cycle scheduling of partitions at the operating system level is supposed for a multiprocessor system. In this paper, we propose a tool for generating such schedules. We use constraint based programming and develop methods and concepts for a combined interactive and automatic partition scheduling system. This paper is also devoted to basic methods and techniques for modeling and solving this partition scheduling problem. Initial application of our partition scheduling tool has proved successful and demonstrated the suitability of the methods used.
|
567 |
Efficiency of difference-list programmingGeske, Ulrich, Goltz, Hans-Joachim January 2010 (has links)
The difference-list technique is described in literature as effective method for extending lists to the right without using calls of append/3. There exist some proposals for automatic transformation of list programs into differencelist programs. However, we are interested in construction of difference-list programs by the programmer, avoiding the need of a transformation step. In [GG09] it was demonstrated, how left-recursive procedures with a dangling call of append/3 can be transformed into right-recursion using the unfolding technique. For simplification of writing difference-list programs using a new cons/2 procedure was introduced. In the present paper, we investigate how efficieny is influenced using cons/2. We measure the efficiency of procedures using accumulator technique, cons/2, DCG’s, and difference lists and compute the resulting speedup in respect to the simple procedure definition using append/3. Four Prolog systems were investigated and we found different behaviour concerning the speedup by difference lists. A result of our investigations is, that an often advice given in the literature for avoiding calls append/3 could not be confirmed in this strong formulation.
|
568 |
AN EFFICIENT SET-BASED APPROACH TO MINING ASSOCIATION RULESHsieh, Yu-Ming 28 July 2000 (has links)
Discovery of {it association rules} is an important problem in the area of data
mining. Given a database of sales transactions, it is desirable to discover
the important associations among items such that the presence of some items
in a transaction will imply the presence of other items in the same
transaction.
Since mining association rules may require to repeatedly scan through a large
transaction database to find different association patterns, the amount of
processing could be huge, and performance improvement is an essential
concern.
Among this problem, how to efficiently {it count large
itemsets} is the major work, where a large itemset is a set of items
appearing in a sufficient number of transactions.
In this thesis, we propose efficient algorithms for mining association
rules based on a high-level set-based approach.
A set-based approach allows a clear expression of what needs to be done
as opposed to specifying exactly how the operations are carried out in a low-level approach, where
a low-level approach means to retrieve one tuple from the database at a time.
The advantage of the set-based approach, like the SETM algorithm,
is simple and stable over the range of parameter values.
However, the SETM algorithm proposed by Houtsma and Swami may generate too many invalid candidate itemsets.
Therefore, in this thesis, we propose a set-based algorithm called SETM*,
which provides the same advantages of the SETM algorithm,
while it avoids the disadvantages of the SETM algorithm.
In the SETM* algorithm, we reduce the size of the candidate database by
modifying the way of constructing it,
where a candidate database is a transaction database formed with candidate
$k$-itemsets.
Then, based on the new way to construct the candidate database in the SETM*
algorithm, we propose SETM*-2K, mbox{SETM*-MaxK} and SETM*-Lmax algorithms.
In the SETM*-2K algorithm, given a $k$, we efficiently construct $L_{k}$
based on $L_{w}$, where $w=2^{lceil log_{2}k
ceil - 1}$,
instead of step by step.
In the SETM*-MaxK algorithm, we efficiently to find the $L_{k}$ based on $L_{w}$,
where $L_{k}
ot= emptyset, L_{k+1}=emptyset$ and $w=2^{lceil log_{2}k
ceil - 1}$,
instead of step by step.
In the SETM*-Lmax algorithm, we use a forward approach to find all maximal large itemsets from $L_{k}$,
and the $k$-itemset is not included in the $k$-subsets of the $j$-itemset,
except $k=MaxK$, where $1 leq k < j leq MaxK$,
$L_{MaxK}
ot= emptyset$ and $L_{MaxK+1}=emptyset$.
We conduct several experiments using different synthetic relational databases.
The simulation results show that the
SETM* algorithm outperforms the SETM algorithm in terms of storage space or the
execution time for all relational database settings.
Moreover, we show that the proposed SETM*-2K
and SETM*-MaxK algorithms also require shorter time to achieve their goals than the SETM or SETM* algorithms.
Furthermore, we also show that the proposed forward approach (SETM*-Lmax)
to find all maximal large itemsets requires shorter time than the backward approach proposed by Agrawal.
|
569 |
Topics in analyzing longitudinal dataJu, Hyunsu 17 February 2005 (has links)
We propose methods for analyzing longitudinal data, obtained in clinical trials
and other applications with repeated measures of responses taken over time. Common
characteristics of longitudinal studies are correlated responses and observations taken
at unequal points in time. The first part of this dissertation examines the justification
of a block bootstrap procedure for the repeated measurement designs, which takes
into account the dependence structure of the data by resampling blocks of adjacent
observations rather than individual data points. In the case of dependent stationary
data, under regular conditions, the approximately studentized or standardized block
bootstrap possesses a higher order of accuracy. With longitudinal data, the second
part of this dissertation shows that the diagonal optimal weights for unbalanced
designs can be made to improve the efficiency of the estimators in terms of mean
squared error criterion. Simulation study is conducted for each of the longitudinal
designs. We will also analyze repeated measurement data set concerning nursing home
residents with multiple sclerosis, which is obtained from a large database termed the
minimum data set (MDS).
|
570 |
Mesoscale ensemble-based data assimilation and parameter estimationAksoy, Altug 01 November 2005 (has links)
The performance of the ensemble Kalman filter (EnKF) in forced, dissipative
flow under imperfect model conditions is investigated through simultaneous state and
parameter estimation where the source of model error is the uncertainty in the model
parameters. Two numerical models with increasing complexity are used with simulated
observations.
For lower complexity, a two-dimensional, nonlinear, hydrostatic, non-rotating,
and incompressible sea breeze model is developed with buoyancy and vorticity as the
prognostic variables. Model resolution is 4 km horizontally and 50 m vertically. The
ensemble size is set at 40. Forcing is maintained through an explicit heating function
with additive stochastic noise. Simulated buoyancy observations on land surface with
40-km spacing are assimilated every 3 hours. Up to six model parameters are
successfully subjected to estimation attempts in various experiments. The overall EnKF
performance in terms of the error statistics is found to be superior to the worst-case scenario (when there is parameter error but no parameter estimation is performed) with
an average error reduction in buoyancy and vorticity of 40% and 46%, respectively, for
the simultaneous estimation of six parameters.
The model chosen to represent the complexity of operational weather forecasting
is the Pennsylvania State University-National Center for Atmospheric Research MM5
model with a 36-km horizontal resolution and 43 vertical layers. The ensemble size for
all experiments is chosen as 40 and a 41st member is generated as the truth with the
same ensemble statistics. Assimilations are performed with a 12-hour interval with
simulated sounding and surface observations of horizontal winds and temperature. Only
single-parameter experiments are performed focusing on a constant inserted into the
code as the multiplier of the vertical eddy mixing coefficient. Estimation experiments
produce very encouraging results and the mean estimated parameter value nicely
converges to the true value exhibiting a satisfactory level of variability.
|
Page generated in 0.0675 seconds