• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 19691
  • 3373
  • 2417
  • 2015
  • 1551
  • 1432
  • 881
  • 406
  • 390
  • 359
  • 297
  • 237
  • 208
  • 208
  • 208
  • Tagged with
  • 38217
  • 12470
  • 9257
  • 7123
  • 6700
  • 5896
  • 5307
  • 5203
  • 4740
  • 3461
  • 3307
  • 2834
  • 2730
  • 2546
  • 2117
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
561

Kato: a plagiarism-detection tool for answer-set programs

Oetsch, Johannes, Schwengerer, Martin, Tompits, Hans January 2010 (has links)
We present the tool Kato which is, to the best of our knowledge, the first tool for plagiarism detection that is directly tailored for answer-set programming (ASP). Kato aims at finding similarities between (segments of) logic programs to help detecting cases of plagiarism. Currently, the tool is realised for DLV programs but it is designed to handle various logic-programming syntax versions. We review basic features and the underlying methodology of the tool.
562

Constraint-based abstraction of a model checker for infinite state systems

Banda, Gourinath, Gallagher, John P. January 2010 (has links)
Abstract interpretation-based model checking provides an approach to verifying properties of infinite-state systems. In practice, most previous work on abstract model checking is either restricted to verifying universal properties, or develops special techniques for temporal logics such as modal transition systems or other dual transition systems. By contrast we apply completely standard techniques for constructing abstract interpretations to the abstraction of a CTL semantic function, without restricting the kind of properties that can be verified. Furthermore we show that this leads directly to implementation of abstract model checking algorithms for abstract domains based on constraints, making use of an SMT solver.
563

Range restriction for general formulas

Brass, Stefan January 2010 (has links)
Deductive databases need general formulas in rule bodies, not only conjuctions of literals. This is well known since the work of Lloyd and Topor about extended logic programming. Of course, formulas must be restricted in such a way that they can be effectively evaluated in finite time, and produce only a finite number of new tuples (in each iteration of the TP-operator: the fixpoint can still be infinite). It is also necessary to respect binding restrictions of built-in predicates: many of these predicates can be executed only when certain arguments are ground. Whereas for standard logic programming rules, questions of safety, allowedness, and range-restriction are relatively easy and well understood, the situation for general formulas is a bit more complicated. We give a syntactic analysis of formulas that guarantees the necessary properties.
564

Transforming imperative algorithms to constraint handling rules

Abdennadher, Slim, Ismail, Haythem, Khoury, Frederick January 2010 (has links)
Different properties of programs, implemented in Constraint Handling Rules (CHR), have already been investigated. Proving these properties in CHR is fairly simpler than proving them in any type of imperative programming language, which triggered the proposal of a methodology to map imperative programs into equivalent CHR. The equivalence of both programs implies that if a property is satisfied for one, then it is satisfied for the other. The mapping methodology could be put to other beneficial uses. One such use is the automatic generation of global constraints, at an attempt to demonstrate the benefits of having a rule-based implementation for constraint solvers.
565

Persistent constraints in constraint handling rules

Betz, Hariolf, Raiser, Frank, Frühwirth, Thom January 2010 (has links)
In the most abstract definition of its operational semantics, the declarative and concurrent programming language CHR is trivially non-terminating for a significant class of programs. Common refinements of this definition, in closing the gap to real-world implementations, compromise on declarativity and/or concurrency. Building on recent work and the notion of persistent constraints, we introduce an operational semantics avoiding trivial non-termination without compromising on its essential features.
566

A tool for generating partition schedules of multiprocessor systems

Goltz, Hans-Joachim, Pieth, Norbert January 2010 (has links)
A deterministic cycle scheduling of partitions at the operating system level is supposed for a multiprocessor system. In this paper, we propose a tool for generating such schedules. We use constraint based programming and develop methods and concepts for a combined interactive and automatic partition scheduling system. This paper is also devoted to basic methods and techniques for modeling and solving this partition scheduling problem. Initial application of our partition scheduling tool has proved successful and demonstrated the suitability of the methods used.
567

Efficiency of difference-list programming

Geske, Ulrich, Goltz, Hans-Joachim January 2010 (has links)
The difference-list technique is described in literature as effective method for extending lists to the right without using calls of append/3. There exist some proposals for automatic transformation of list programs into differencelist programs. However, we are interested in construction of difference-list programs by the programmer, avoiding the need of a transformation step. In [GG09] it was demonstrated, how left-recursive procedures with a dangling call of append/3 can be transformed into right-recursion using the unfolding technique. For simplification of writing difference-list programs using a new cons/2 procedure was introduced. In the present paper, we investigate how efficieny is influenced using cons/2. We measure the efficiency of procedures using accumulator technique, cons/2, DCG’s, and difference lists and compute the resulting speedup in respect to the simple procedure definition using append/3. Four Prolog systems were investigated and we found different behaviour concerning the speedup by difference lists. A result of our investigations is, that an often advice given in the literature for avoiding calls append/3 could not be confirmed in this strong formulation.
568

AN EFFICIENT SET-BASED APPROACH TO MINING ASSOCIATION RULES

Hsieh, Yu-Ming 28 July 2000 (has links)
Discovery of {it association rules} is an important problem in the area of data mining. Given a database of sales transactions, it is desirable to discover the important associations among items such that the presence of some items in a transaction will imply the presence of other items in the same transaction. Since mining association rules may require to repeatedly scan through a large transaction database to find different association patterns, the amount of processing could be huge, and performance improvement is an essential concern. Among this problem, how to efficiently {it count large itemsets} is the major work, where a large itemset is a set of items appearing in a sufficient number of transactions. In this thesis, we propose efficient algorithms for mining association rules based on a high-level set-based approach. A set-based approach allows a clear expression of what needs to be done as opposed to specifying exactly how the operations are carried out in a low-level approach, where a low-level approach means to retrieve one tuple from the database at a time. The advantage of the set-based approach, like the SETM algorithm, is simple and stable over the range of parameter values. However, the SETM algorithm proposed by Houtsma and Swami may generate too many invalid candidate itemsets. Therefore, in this thesis, we propose a set-based algorithm called SETM*, which provides the same advantages of the SETM algorithm, while it avoids the disadvantages of the SETM algorithm. In the SETM* algorithm, we reduce the size of the candidate database by modifying the way of constructing it, where a candidate database is a transaction database formed with candidate $k$-itemsets. Then, based on the new way to construct the candidate database in the SETM* algorithm, we propose SETM*-2K, mbox{SETM*-MaxK} and SETM*-Lmax algorithms. In the SETM*-2K algorithm, given a $k$, we efficiently construct $L_{k}$ based on $L_{w}$, where $w=2^{lceil log_{2}k ceil - 1}$, instead of step by step. In the SETM*-MaxK algorithm, we efficiently to find the $L_{k}$ based on $L_{w}$, where $L_{k} ot= emptyset, L_{k+1}=emptyset$ and $w=2^{lceil log_{2}k ceil - 1}$, instead of step by step. In the SETM*-Lmax algorithm, we use a forward approach to find all maximal large itemsets from $L_{k}$, and the $k$-itemset is not included in the $k$-subsets of the $j$-itemset, except $k=MaxK$, where $1 leq k < j leq MaxK$, $L_{MaxK} ot= emptyset$ and $L_{MaxK+1}=emptyset$. We conduct several experiments using different synthetic relational databases. The simulation results show that the SETM* algorithm outperforms the SETM algorithm in terms of storage space or the execution time for all relational database settings. Moreover, we show that the proposed SETM*-2K and SETM*-MaxK algorithms also require shorter time to achieve their goals than the SETM or SETM* algorithms. Furthermore, we also show that the proposed forward approach (SETM*-Lmax) to find all maximal large itemsets requires shorter time than the backward approach proposed by Agrawal.
569

Topics in analyzing longitudinal data

Ju, Hyunsu 17 February 2005 (has links)
We propose methods for analyzing longitudinal data, obtained in clinical trials and other applications with repeated measures of responses taken over time. Common characteristics of longitudinal studies are correlated responses and observations taken at unequal points in time. The first part of this dissertation examines the justification of a block bootstrap procedure for the repeated measurement designs, which takes into account the dependence structure of the data by resampling blocks of adjacent observations rather than individual data points. In the case of dependent stationary data, under regular conditions, the approximately studentized or standardized block bootstrap possesses a higher order of accuracy. With longitudinal data, the second part of this dissertation shows that the diagonal optimal weights for unbalanced designs can be made to improve the efficiency of the estimators in terms of mean squared error criterion. Simulation study is conducted for each of the longitudinal designs. We will also analyze repeated measurement data set concerning nursing home residents with multiple sclerosis, which is obtained from a large database termed the minimum data set (MDS).
570

Mesoscale ensemble-based data assimilation and parameter estimation

Aksoy, Altug 01 November 2005 (has links)
The performance of the ensemble Kalman filter (EnKF) in forced, dissipative flow under imperfect model conditions is investigated through simultaneous state and parameter estimation where the source of model error is the uncertainty in the model parameters. Two numerical models with increasing complexity are used with simulated observations. For lower complexity, a two-dimensional, nonlinear, hydrostatic, non-rotating, and incompressible sea breeze model is developed with buoyancy and vorticity as the prognostic variables. Model resolution is 4 km horizontally and 50 m vertically. The ensemble size is set at 40. Forcing is maintained through an explicit heating function with additive stochastic noise. Simulated buoyancy observations on land surface with 40-km spacing are assimilated every 3 hours. Up to six model parameters are successfully subjected to estimation attempts in various experiments. The overall EnKF performance in terms of the error statistics is found to be superior to the worst-case scenario (when there is parameter error but no parameter estimation is performed) with an average error reduction in buoyancy and vorticity of 40% and 46%, respectively, for the simultaneous estimation of six parameters. The model chosen to represent the complexity of operational weather forecasting is the Pennsylvania State University-National Center for Atmospheric Research MM5 model with a 36-km horizontal resolution and 43 vertical layers. The ensemble size for all experiments is chosen as 40 and a 41st member is generated as the truth with the same ensemble statistics. Assimilations are performed with a 12-hour interval with simulated sounding and surface observations of horizontal winds and temperature. Only single-parameter experiments are performed focusing on a constant inserted into the code as the multiplier of the vertical eddy mixing coefficient. Estimation experiments produce very encouraging results and the mean estimated parameter value nicely converges to the true value exhibiting a satisfactory level of variability.

Page generated in 0.0675 seconds