• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 367
  • 47
  • 39
  • 15
  • 9
  • 9
  • 9
  • 9
  • 9
  • 9
  • 9
  • 8
  • 7
  • 6
  • 5
  • Tagged with
  • 637
  • 164
  • 154
  • 152
  • 121
  • 118
  • 115
  • 78
  • 75
  • 72
  • 65
  • 64
  • 63
  • 60
  • 56
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
251

Nonparametric item response modeling for identifying differential item functioning in the moderate-to-small-scale testing context

Witarsa, Petronilla Murlita 11 1900 (has links)
Differential item functioning (DIF) can occur across age, gender, ethnic, and/or linguistic groups of examinee populations. Therefore, whenever there is more than one group of examinees involved in a test, a possibility of DIF exists. It is important to detect items with DIF with accurate and powerful statistical methods. While finding a proper DIP method is essential, until now most of the available methods have been dominated by applications to large scale testing contexts. Since the early 1990s, Ramsay has developed a nonparametric item response methodology and computer software, TestGraf (Ramsay, 2000). The nonparametric item response theory (IRT) method requires fewer examinees and items than other item response theory methods and was also designed to detect DIF. However, nonparametric IRT's Type I error rate for DIF detection had not been investigated. The present study investigated the Type I error rate of the nonparametric IRT DIF detection method, when applied to moderate-to-small-scale testing context wherein there were 500 or fewer examinees in a group. In addition, the Mantel-Haenszel (MH) DIF detection method was included. A three-parameter logistic item response model was used to generate data for the two population groups. Each population corresponded to a test of 40 items. Item statistics for the first 34 non-DIF items were randomly chosen from the mathematics test of the 1999 TEVISS (Third International Mathematics and Science Study) for grade eight, whereas item statistics for the last six studied items were adopted from the DIF items used in the study of Muniz, Hambleton, and Xing (2001). These six items were the focus of this study. / Education, Faculty of / Educational and Counselling Psychology, and Special Education (ECPS), Department of / Graduate
252

Eksamenskryfvaardighede by eerstejaarstudente met spesifieke verwysing na meervoudige keusevrae

Van den Berg, Hester Regina 17 November 2014 (has links)
M.A.(Psychology) / Please refer to full text to view abstract
253

The effectiveness of mindfulness meditation on reducing test-taking anxiety

Griffin, Jeffrey Michael 01 January 1994 (has links)
No description available.
254

Web based entry level mathematics test

Baek, Okbun 01 January 2007 (has links)
The primary purpose of the project is to develop a web site where students can practice entry level mathematics questions.
255

A Cognitively Diagnostic Modeling Approach to Diagnosing Misconceptions and Subskills

Elbulok, Musa January 2021 (has links)
The objective of the present project was to propose a new methodology for measuring misconceptions and subskills simultaneously using diagnostic information available from incorrect alternatives in multiple-choice tests designed for that purpose. Misconceptions are systematic and persistent errors that represent a learned intentional incorrect response (Brown & VanLehn, 1980; Ozkan & Ozkan, 2012). In prior research, Lee and Corter (2011) found that classification accuracy for their Bayesian Network misconception diagnosis models improved when latent higher-order subskills and specific wrong answers were included. Here, these contributions are adapted to a cognitively diagnostic measurement approach using the multiple-choice Deterministic Inputs Noisy “And” Gate (MC-DINA) model, first developed by de la Torre (2009b), by specifying dependencies between attributes to measure latent misconceptions and subskills simultaneously. A simulation study was conducted employing the proposed methodology (referred to as MC-DINA-H) across sample sizes (500, 1000, 2,000, and 5,000 examinees) and test lengths (15, 30, and 60 items) conditions. Eight attributes (4 misconceptions and 4 subskills) were included in the main simulation study. Attribute classification accuracy of the MC-DINA-H was compared to four less complex models and was found to more accurately classify attributes only when the attributes were relatively frequently required by multiple-choice options in the diagnostic assessment. The findings suggest that each attribute should be required by at least 15-20 percent of options in the diagnostic assessment.
256

Test Materials for Fundamental Elements in Clerical and Secreterial Subject Matter

Evans, Louise 08 1900 (has links)
This thesis presents an overview and analysis of tests conducted to measure the quality of clerical and secreterial skills obtained by commerce education students and concludes with suggestions for future testing.
257

Exploring a Generalizable Machine Learned Solution for Early Prediction of Student At-Risk Status

Coleman, Chad January 2021 (has links)
Determining which students are at-risk of poorer outcomes -- such as dropping out, failing classes, or decreasing standardized examination scores -- has become an important area of both research and practice in K-12 education. The models produced from this type of predictive modeling research are increasingly used by high schools in Early Warning Systems to identify which students are at risk and intervene to support better outcomes. It has become common practice to re-build and validate these detectors, district-by-district, due to different data semantics and various risk factors for students in different districts. As these detectors become more widely used, however, a new challenge emerges in applying these detectors across a broad spectrum of school districts with varying availability of past student data. Some districts have insufficient high-quality past data for building an effective detector. Novel approaches that can address the complex data challenges a new district presents are critical for advancing the field. Using an ensemble-based algorithm, I develop a modeling approach that can generate a useful model for a previously unseen district. During the ensembling process, my approach, District Similarity Ensemble Extrapolation (DSEE), weights districts that are more similar to the Target district more strongly during ensembling than less similar districts. Using this approach, I can predict student-at-risk status effectively for unseen districts, across a range of grade ranges, and achieve prediction goodness but ultimately fails to perform better than the previously published Knowles (2015) and Bowers (2012) EWS models proposed for use across districts.
258

Iowa placement examinations

Stoddard, George Dinsmore 01 January 1925 (has links)
No description available.
259

A comparison of academic achievement in modular and traditional scheduled high schools on province of Quebec high school leaving examinations /

Schuddeboom, James Frederick January 1973 (has links)
No description available.
260

A study of the comprehensive examination administered to graduate students in the Department of Education, Massachusetts State College, 1939.

Fitzgerald, John A. 01 January 1940 (has links) (PDF)
No description available.

Page generated in 0.3992 seconds