• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 138
  • 20
  • 17
  • 10
  • 6
  • 5
  • 4
  • 3
  • 3
  • 2
  • 2
  • 2
  • 2
  • 2
  • 1
  • Tagged with
  • 259
  • 44
  • 39
  • 37
  • 32
  • 26
  • 22
  • 17
  • 17
  • 17
  • 17
  • 17
  • 16
  • 16
  • 16
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
71

ACT Score Declines: Looking for the Source

Atkin, Thomas Edward 01 May 1979 (has links)
During the last 15 years there has been a well documented decline in achievement test scores. Declines have been documented on both the American College Testing Program (ACT) and the Scholastic Aptitude Test (SAT). An increased interest in t he score declines and their possible explanations has also taken place in the last 10 years. The explanation for the declines fit under four main headings: (1) problems with the tests, (2) changes in the testing population, (3) changes in society, and (4) changes in the schools. The purpose of this study was to begin with the test itself and check for possible score drift between two forms of the ACT, a 1977 ACT form and a form 5 to 7 years older than that. The sample population consisted of 242 juniors at Logan High School, Logan, Utah, and 153 juniors at Skyview High School, Smithfield, Utah, during the 1976-77 school year. These samples represent 83% of the Logan High junior population and 50% of the Skyview High junior class. All subjects took the two forms of the ACT mentioned above. In analyzing the data from the major samples with a two-tailed t test, it was found that there was a significant difference in the two tests. The largest difference was found on the mathematics subtest; the difference being that students scored higher on the older forms, not only on the mathematics subtest, but on the other subtests, also. A two-way analysis of variance with repeated measures was used to check for variance due to sex of subjects and form of test and their interractions. The findings were consistent with those of the t test. It was found that the mathematics scores varied due to the form of the test. The only sex difference was that females' standard scores were higher than males' on the English subtests of both ACT forms. This study found a significant difference between an older ACT form and a 1977 ACT form as measured by subjects' scores on both. The findings of this research indicate that the test itself may be a partial cause of the current ACT score declines.
72

A study of the comparative validity of a scholastic aptitude test and an achievement test in predicting school success

Moriarty, Helen E. 01 January 1933 (has links) (PDF)
No description available.
73

I wouldn't steal a car but I would download one if I could - En kvalitativ undersökning av fildelning ur ett kriminologiskt perspektiv

Totting, Jakob January 2010 (has links)
Illegal fildelning är ett världsomfattande problem som har skapat heta debatter i många länder. I Sverige är den illegala fildelningen ett vanligt förekommande fenomen. För att komma till rätta med denna verksamhet har vår egen upphovsrättslag uppdaterats och den omdiskuterade IPRED-lagen införts. Trots detta är företeelsen illegal fildelning fortfarande vanligt förekommande i dagens Sverige. Den här studien har haft som syfte attundersöka varför personer fortfarande fildelar samt vad som skulle få dem att sluta med dennaverksamhet. Resultatet av studien visar bl a att man inte upplever risken för upptäckt och straff som stor. Gratis tillgänglighet av digitalt material på nätet i kombination med dagens snabba och smidiga bredbandsuppkopplingar inbjuder också till fortsatt fildelning. Två kriminologiska teorier; Situationell handlingsteori och Rutinaktivitetsteorin framkommerur denna kvalitativa studie som förklaring till fildelarnas handlande. Utifrån fildelarna självaskulle illegal fildelning kunna fås att upphöra om det infördes fler lagliga alternativ för ”streaming” (strömning) av media eller om kostnaderna för digitalt material på Internet sänktes till en lägre och rimligare nivå. Studien visar således varför den illegala fildelningen fortfarande är vanligt förekommande samt redovisar också några alternativ till den illegala fildelningen. En handling som har gjorts straffbar och därmed en stor del av vår befolkning kriminell.
74

APPLICATIONS OF SATISFIABILITY IN SYNTHESIS OF RECONFIGURABLE COMPUTERS

SIVA, SUBRAMANYAN D. 11 June 2002 (has links)
No description available.
75

A Formal Approach to Concurrent Error Detection in FPGA LUTs

Bergstra, Jameson P. 10 1900 (has links)
In this thesis we discuss a formal approach to the design of concurrent error detection (CED) logic in field-programmable gate arrays (FPGAs). Single event upsets (SEUs) occurring in look-up table (LUT) configuration bits are considered as the fault model. Our approach involves representing the LUT network of the design implemented in the FPGA with constraints to model the presence of SEUs as a boolean formula in conjunctive normal form. A quantified boolean formula (QBF) based approach to designing CED logic based on parity check codes is found to be infeasible for designs of a realistic size. It is shown that a satisfiability (SAT) solver can be used to find variable assignments that indicate which circuit outputs can be corrupted by upset events in the specified fault model. An algorithm is presented to automatically generate a parity check code, which will identify with one clock cycle detection latency a malfunction caused by an SEU. The resulting parity check logic can be verified using a SAT solver and it is shown to require fewer LUT resources than duplication for most circuits. / Master of Applied Science (MASc)
76

The Relationship Between DIBELS Oral Reading Fluency Scores and Reading Scores on High-Stakes Assessments

Curry, Alicia Lenise 30 April 2011 (has links)
Early identification of children with reading problems is a vital element of academic success at all grade levels. Thus, it is crucial that educators select and implement efficient reading assessments and procedures. The purpose of this study was to investigate third grade archival data of students over a 2-year period at one school to determine if a relationship existed between the DIBELS 3rd grade oral reading fluency scores and the reading scores on Stanford Achievement Test and the reading scores on the Alabama Reading and Math Test. Archival data from 80 third grade African American students who were assessed with all three assessments was used to conduct this study. Pearson and Spearman statistical tests were performed on the data to examine the relationship between DIBELS (oral reading fluency) and the reading section of the Stanford Achievement Test and the reading section of the Alabama Reading and Math Test. The results indicated a relationship existed between DIBELS oral reading fluency scores and the reading scores on Stanford Achievement Test and the reading scores on the Alabama Reading and Math Test during the 2009-2010 school years.
77

Design Verification for Sequential Systems at Various Abstraction Levels

Zhang, Liang 31 January 2005 (has links)
With the ever increasing complexity of digital systems, functional verification has become a daunting task to circuit designers. Functional verification alone often surpasses 70% of the total development cost and the situation has been projected to continue to worsen. The most critical limitations of existing techniques are the capacity issue and the run-time issue. This dissertation addresses the functional verification problem using a unified approach, which utilizes different core algorithms at various abstraction levels. At the logic level, we focus on incorporating a set of novel ideas to existing formal verification approaches. First, we present a number of powerful optimizations to improve the performance and capacity of a typical SAT-based bounded model checking framework. Secondly, we present a novel method for performing dynamic abstraction within a framework for abstraction-refinement based model checking. Experiments on a wide range of industrial designs have shown that the proposed optimizations consistently provide between 1-2 orders of magnitude speedup and can be extremely useful in enhancing the efficacy of existing formal verification algorithms. At the register transfer level, where the formal verification is less likely to succeed, we developed an efficient ATPG-based validation framework, which leverages the high-level circuit information and an improved observability-enhanced coverage to generate high quality validation sequences. Experiments show that our approach is able to generate high quality validation vectors, which achieve both high tag coverage and high bug coverage with extremely low computational cost. / Ph. D.
78

Exploring Abstraction Techniques for Scalable Bit-Precise Verification of Embedded Software

He, Nannan 01 June 2009 (has links)
Conventional testing has become inadequate to satisfy rigorous reliability requirements of embedded software that is playing an increasingly important role in many safety critical applications. Automatic formal verification is a viable avenue for ensuring the reliability of such software. Recently, more and more formal verification techniques have begun modeling a non-Boolean data variable as a bit-vector with bounded width (i.e. a vector of multiple bits like 32- or 64- bits) to implement bit-precise verification. One major challenge in the scalable application of such bit-precise verification on real-world embedded software is that the state space for verification can be intractably large. In this dissertation, several abstraction techniques are explored to deal with this scalability challenge in the bit-precise verification of embedded software. First, we propose a tight integration of program slicing, which is an important static program analysis technique, with bounded model checking (BMC). While many software verification tools apply program slicing as a separate preprocessing step, we integrate slicing operations into our model construction and reduction process and enhance them with compilation optimization techniques to compute accurate program slices. We also apply a proof-based abstraction-refinement framework to further remove those program segments irrelevant to the property being verified. Next, we present a method of using symbolic simulation for scalable formal verification. The simulation involves distinguishing X as symbolic values to abstract concrete variables' values. Also, the method embeds this symbolic simulation in a counterexample-guided abstraction-refinement framework to automatically construct and verify an abstract model, which has a smaller state space than that of the original concrete program. This dissertation also presents our efforts on using two common testability metrics — controllability metric (CM) and observability metric (OM) — as the high-level structural guidance for scalable bit-precise verification. A new abstraction approach is proposed based on the concept of under- and over-approximation to efficiently solve bit-vector formulas generated from embedded software verification instances. These instances include both complicated arithmetic computations and intensive control structures. Our approach applies CM and OM to assist the abstraction refinement procedure in two ways: (1) it uses CM and OM to guide the construction of a simple under-approximate model, which includes only a subset of execution paths in a verification instance, so that a counterexample that refutes the instance can be obtained with reduced effort, and (2) in order to reduce the cost of using proof-based refinement alone, it uses OM heuristics to guide the restoration of additional verification-relevant formula constraints with low computational cost for refinement. Experiments show a significant reduction of the solving time compared to state-of-the-art solvers for the bit-vector arithmetic. This dissertation finally proposes an efficient algorithm to discover non-uniform encoding widths of individual variables in the verification model, which may be smaller than their original modeling width but sufficient for the verification. Our algorithm distinguishes itself from existing approaches in that it is path-oriented; it takes advantage of CM and OM values to guide the computation of the initial, non-uniform encoding widths, and the effective adjustment of these widths along different paths, until the property is verified. It can restrict the search from those paths that are deemed less favorable or have been searched in previous steps, thus simplifying the problem. Experiments demonstrate that our algorithm can significantly speed up the verification especially in searching for a counterexample that violates the property under verification. / Ph. D.
79

Search-space Aware Learning Techniques for Unbounded Model Checking and Path Delay Testing

Chandrasekar, Kameshwar 24 April 2006 (has links)
The increasing complexity of VLSI designs, in recent years, poses serious challenges while ensuring the correctness of large designs for functionality and timing. In this dissertation, we target two related problems in Design Verification and Testing: Unbounded Model Checking and Path Delay Fault Testing, that commonly suffer from extremely large memory requirements. We propose efficient representations and intelligent learning techniques that reason on the problem structure and take advantage of the repeated search space, thereby alleviating the memory required and time taken to solve these problems. In this dissertation, we exploit Automatic Test Pattern Generation (ATPG) for Unbounded Model Checking (UMC). In order to perform unbounded model checking, we need the core image / preimage computation engines that perform forward / backward reachability analysis. First, we develop an ATPG engine, with search-space aware learning, that computes ``all solutions" for a given target objective and stores it as a decision diagram. We propose efficient decision selection heuristics and derive a suitable cut-set metric to quickly obtain a compact solution set. The solution set that is obtained, with the initial state set as the objective, represents the one-cycle preimage. In order to use the preimage state set as the objective in the subsequent iterations, we propose efficient techniques to convert a decision diagram into clauses/circuit. We propose a node-based conversion scheme that derives the functionality of each node in the decision diagram. The proposed scheme contains the size of the state set and helps to iteratively compute the preimage for many cycles until a fixed point / desired state is reached. Further, we gear the ATPG engine to directly compute the circuit cofactors, rather than individual solutions. The circuit cofactors contain a large number of solutions and hence capture a larger solution space. We also propose efficient learning techniques to prune the cofactor space and accelerate preimage computation. Then, we develop an exclusive image computation procedure that branches on the combinational inputs of the circuit and projects the values on the next state flip-flops as the image. We perform learning on the input solution space and incrementally store the image obtained as a decision diagram. We consistently show, with our experimental results, that our techniques are better than the existing techniques in terms of both performance and capacity. In the case of delay testing, we consider the test generation for path delay fault (PDF) model, which is the most accurate in characterizing the cumulative effect of distributed delays along each path in a circuit. The main bottle-neck in the ATPG for PDFs is the exponential number of paths in a circuit. In this work, we use the circuit information to analyze the common segments shared by different paths in a circuit. Based on the common sensitization constraints, we propose to identify the ``untestable core of segments" that cannot be sensitized together. We use these segments to identify the conflict search space for a huge number of untestable path delay faults apriori and prune them on-the-fly during test generation. Experimental results show that a huge number of untestable path delay faults are identified and it helps to accelerate test generation. / Ph. D.
80

Strategies for SAT-Based Formal Verification

Vimjam, Vishnu Chaithanya 13 February 2007 (has links)
Verification of digital hardware designs is becoming an increasingly complex task as the designs are incorporating more functionality, becoming complex and growing larger in size. Today, verification remains a bottleneck in meeting time-to-market requirements and consumes more than 70% of the overall design-costs. Traditionally, verification has been done using simulation-based approaches, where a set of appropriate test-stimuli is used by the designer. As the designs become more complex, however, simulation-based techniques often fail to capture corner-case errors. Furthermore, unless exhaustively tested, these approaches do not guarantee the correctness of a system with respect to its specifications. As a consequence, formal methods for design verification have been sought after. In formal verification, the conformance of a design to a given set of specifications is proven mathematically, thereby leaving no room for unexplored search spaces. Despite the exponential time/memory complexities often involved within the formal approaches, they have shown promise in capturing subtle bugs, which were missed otherwise. In this dissertation, we focus on Boolean Satisfiability (SAT) based formal verification, which has gained tremendous importance in the recent past. Importantly, SAT-based approaches often alleviate the memory explosion problem, which had been a bottleneck of the traditional symbolic (Binary Decision Diagram based) approaches. In SAT-based techniques, the set of verification tasks are converted into a set of Boolean formulae, which are checked for satisfiability using a SAT solver. These problems are often NP-complete and are prone to an explosion in the required run-time. To overcome this, we propose novel strategies which utilize both structural and logical information of a sequential circuit. In particular, we devise techniques to extract non-trivial invariants of a design, strengthen properties such that they can be proven faster and interleave bounded reachability analysis with bounded model checking. We provide the necessary algorithms and implementation details in order to automate the proposed techniques. Experiments conducted on a variety of benchmark circuits show that orders of magnitude improvement in overall run-times can be achieved via our techniques compared to the existing state-of-the-art SAT-based approaches. / Ph. D.

Page generated in 0.3565 seconds