Spelling suggestions: "subject:"educationization -- tests anda remeasurements"" "subject:"educationization -- tests anda aremeasurements""
91 |
Distribution and power of selected item bias indices: A Monte Carlo study.Ibrahim, Abdul K. January 1992 (has links)
This study examines the following DIF procedures--Transformed Item Difficulty (TID), Full Chi-Square, Mantel-Haenszel chi-square, Mantel-Haenszel delta, Logistic Regression, SOS2, SOS4, and Lord's chi-square under three sample sizes, two test lengths, four cases of item discrimination arrangement, and three item difficulty levels. The study is in two parts: The first part examines the distributions of the indices under null (no bias) conditions. The second part deals with the power of the procedures to detect known bias in simulated test data. Agreements among procedures are also addressed. Lord's chi-square certainly appears to perform very well. Its detection rates were very good, and its percentiles were not affected by discrimination level or test length. In retrospect, one would like to know how well it might do at smaller sample sizes. When the tabled values were used, it performed equally well in detecting bias and improved in reducing false positive rates. Of the other indices, the Mantel-Haenszel and the logistic regression indices seemed the best. Camilli chi-square had a number of problems. Its tabled values were not at all useful for detection of bias. The TID was somewhat better but does not have a significance test associated with it. One would need to rely on baseline studies, if one were to use it. For uniform bias either Mantel-Haenszel chi-square or logistic regression would be recommended, while for nonuniform bias logistic regression would be appropriate. It is interesting to note that Lord's chi-square was effective for detecting either kinds of bias. We have been told that sample size is related to chi-square values. For each of the chi square indices the observed values were considerably lower than tabled values. Of course, these were conditions where no bias was present except that which might be randomly induced in data generation. Perhaps it is those instances where bias is truly present that larger sample sizes allow us to more easily identify biased items. Certainly the proportions of biased items detected was greater for large sample sizes for Camilli chi-square, Mantel-Haenszel chi-square, and logistic regression chi-squares.
|
92 |
Étude corrélationnelle des liens entre l'anticipation, la préparation et l'autoévaluation, et le résultat à un examen de rendement scolaire.Fournier, Charles. January 1991 (has links)
Abstract Not Available.
|
93 |
Assessing test dimensionality using two approximate chi-square statistics.De Champlain, André F. January 1992 (has links)
Abstract Not Available.
|
94 |
An empirical study of the consistency of differential item functioning detection.Brown, Paulette C. January 1992 (has links)
Total test scores of examinees on any given standardized test are used to provide reliable and objective information regarding the overall performance of the test takers. When the probability of successfully responding to a test item is not the same for examinees at the same ability levels, but from different groups, the item functions differentially in favour of one group over the other group. This type of problem, defined as differential item functioning (DIF), creates a disadvantage for members of certain subgroups of test takers. Test items need to be accurate and valid measures for all groups because test results may be used to make significant decisions which may have an impact on the future opportunities available to test takers. Thus, DIF is an issue of concern in the field of educational measurement. The purpose of this study was to investigate how well the Mantel-Haenszel (MH) and logistic regression (LR) procedures perform in the identification of items that function differentially across gender groups and regional groups. Research questions to be answered by this study were concerned with three issues: (1) the detection rates for DIF items and items which did not exhibit DIF, (2) the agreement for the MH and LR methods in the detection of DIF items, and (3) the effectiveness of these indices across sample size and over replications. (Abstract shortened by UMI.)
|
95 |
Detecting DIF in polytomous item responses.Tian, Fang. January 1999 (has links)
Use of polytomously scored items in educational tests is becoming increasingly common with new types of assessments such as performance assessment It is believed that performance assessment provides more equitable approaches to testing than traditional multiple-choice tests. However, some forms of performance assessment may be more likely than conventional tests to be influenced by construct-irrelevant factors, resulting in differential item functioning (DIF). Several methods have been proposed for DIF assessment with polytomous item responses, such as the Mantel procedure (Mantel, 1963), the generalized Mantel-Haenszel procedure (GMH) (Mantel & Haenszel, 1959), the polytomous extension of the standardization approach (STND) (Potenza & Dorans, 1995), the polytomous extension of SIBTEST (SIBTEST) (Chang, Mazzeo, & Roussos, 1996), and logistic discriminant function analysis (LDFA) (Miller & Spray, 1993). The purpose of this study was to investigate and compare the performance of the Mantel, GMH, STND, SIBTEST, and LDFA (LDFA-Uniform and LDFA-Nonuniform) procedures in detecting DIF with polytomous item responses under a variety of conditions. A simulation study was conducted to evaluate their Type I error and power in detecting DIF when the properties of the items were known. The factors considered were group ability distribution differences (three levels), test length (two levels), sample size (three levels), sample size ratio (two levels), item discrimination (null DIF, constant uniform DIF, and balanced uniform DIF) or item-discrimination parameter difference between reference and focal groups (nonuniform DIF) (three levels), and DIF conditions (four levels). Based on the findings of this study, all procedures had good and comparable Type I error control when groups were equal in ability or when groups had a small ability difference and item discrimination was not high or when groups had a large ability difference and item discrimination was low. Both LDFA-Nonuniform and GMH can be used to detect nonuniform DIF with LDFA-Nonuniform being more powerful when two groups had equal ability distributions and GMH being more powerful when two groups had different ability distributions. Finally, results from this study indicated that the LDFA procedure is preferable over the other procedures. (Abstract shortened by UMI.)
|
96 |
Assessing the performance of the approximate chi-square and Stout's T statistics with different test structures.Pang, Xiao L. January 1999 (has links)
The assessment of dimensionality underlying the responses to a set of test items is a fundamental issue in applying correct IRT models to examinee ability estimation and test result interpretation. Currently, three assessment methods have been shown to be particularly promising: the original Stout's T (Stout's T1), the refined Stout's T (Stout's T2) (Nandakumar, 1987, 1993) based on Stout's essential unidimensional assumption (Stout, 1987), and the Approximate chi2 (De Champlain & Gessaroli, 1992) derived from McDonald's nonlinear factor analysis based on the weak principle of local independence (McDonald, 1981). However, the three indices have only been tested under limited research conditions. The purpose of this study was to assess and compare the Type I error rates and power of the Approximate chi2, Stout's T1, and Stouts T2 in assessing dimensionality of a set of item responses. The variables used in the Type I error study were test length (L) (40 and 80 items), sample size (N) (500, 1,000, and 2,000), and item discrimination (a) (.7, 1.0, and .14). A 2 x 3 x 3 design was created. For each cell of the design, 100 replications were carried out. In the power study, in addition to the three variables used in the Type I error study, different test structure (two dimensional simple test structure and two dimensional complex test structure) and dimension correlation (r = .0, .4, .57, and .7) were used. For both simple and complex test structure, a dimension ratio of 3:1 was set. Similarly, 100 replications were carried out for each combination of the conditions. A total of 14,400 data sets were generated. A according to the results obtained in this study, each index possesses certain advantages and drawbacks. The Approximate chi2 had good Type I error of zero over all conditions and excellent power with two dimensional simple test structure. Stout's T1 and Stout's T2, on the other hand, had higher Type I error rates than the Approximate chi 2, ranging from zero to 12, excellent power with two dimensional simple test structure, and better power than the Approximate chi2 with two dimensional complex test structure. However, Stouts T1 and Stouts T2 have to be used with great caution given the unsatisfactory power shown with two dimensional complex test structure in many cases. (Abstract shortened by UMI.)
|
97 |
Identification of gifted students using a neuropsychophysiological paradigm: An exploratory study.Shaffer, Dianna. January 1999 (has links)
The development of an accurate assessment paradigm that identifies students as capable of high performance in a variety of areas is urgently needed in education. Current strategies tend to identify only those students who are intellectually or academically gifted. Gifted students who possess other strengths or pupils who exhibit atypical behavior are often overlooked or misidentified in the process. A few educational theorists/practitioners, believe that current practice should be supplemented by some of the testing protocols typically associated with the field of neuropsychophysiology (i.e. neurology, psychology, physiology). This study develops a neuropsychophysiological paradigm that differentiates high average, gifted, and gifted students with perceived behavior problems. Sixty-six students, 10--12 years of age, participated in the study. After being screened for incidents of psychopathology, each student completed 14 timed/untimed relaxation and performance conditions while being monitored by EEG and biofeedback technology. The data were analysed quantitatively using three non-parametric tests. The results of the data analysis suggest that high average, gifted, and gifted students with perceived behavior problems can be differentiated using a neuropsychophysiological paradigm. Foremost in the conclusions is the suggestion that gifted students may excel in two domains intellectual and psychophysiological, and that gifted students with perceived behavior problems may be a legitimate subgroup of the larger population.
|
98 |
Étude des éléments d'évaluation de l'enseignement clinique en soins infirmiers au niveau collégial.Pharand, Denyse. January 1999 (has links)
Le but premier de la présente étude est de préciser l'intention du processus d'évaluation de l'enseignement clinique, c'est-à-dire les éléments pouvant constituer l'objet de l'évaluation et ce, à partir des perceptions des professeures et des étudiantes en soins infirmiers au niveau collégial. Les composantes du processus de l'enseignement en milieu clinique et la notion du rôle sont également étudiés à titre complémentaire.
Le cadre conceptuel de cette étude est basé sur la complémentarité des travaux de recherche sur l'enseignement clinique dans le domaine des sciences infirmières; les uns touchant la définition du processus de l'enseignement clinique, les autres portant sur l'évaluation de cet aspect de l'enseignement. La recherche s'est déroulée auprès de six professeures et six étudiantes. Dans un premier temps, les 12 participantes ont été observées directement dans leur milieu d'enseignement clinique, par la suite chacune d'entre elles a été rencontrée en entrevue. Afin de vérifier l'exactitude de l'analyse des données, deux démarches de validation ont été effectuées. Une première validation, auprès de deux professeures de soins infirmiers non impliquées dans l'étude, où ces dernières ont étudié les protocoles d'entrevue codés de même que les cartes concepts correspondantes. La seconde démarche de validation a été faite auprès de chacune des participantes.
Les résultats indiquent que la relation privilégiée entre la professeure et ses étudiantes se situe au coeur du processus de l'enseignement clinique. Cette conception du processus de l'enseignement clinique introduit les éléments d'évaluation de l'enseignement clinique, car la professeure qui est imputable du maintien de la qualité de la relation avec ses étudiantes, se doit de développer cinq compétences principales. Ces compétences sont (a) humaine, (b) pédagogique, (c) technique, (d) professionnelle infirmière et (e) organisationnelle. Ces compétences principales sont concrétisées par des compétences sateIlites, à leur tour explicitées par des indicateurs.
La présente recherche, en plus de jeter les premiers jalons pour l'élaboration des instruments d'évaluation, devrait apporter tant aux professeures qui enseignent en soins infirmiers, qu'aux administrateurs des institutions d'enseignement. Pour les premières, les compétences identifiées comme éléments d'évaluation offrent un guide précieux pour bien se préparer à leur fonction d'enseignement en milieu clinique. Pour les seconds, les résultats de cette étude donnent un cadre de référence précis pour initier un processus d'évaluation de l'enseignement en milieu clinique dont l'intention pourrait tre formative ou sommative. L'approche qualitative et analytique utilisée dans cette recherche pourrait également servir de modèle générique pour développer le construit du processus d'évaluation de l'enseignement pratique pour d'autres professions. (Abstract shortened by UMI.)
|
99 |
A comparison of the sample invariance of item statistics from the classical test model, item response model, and structural equation model: A case study of real response data.Breithaupt, Krista J. January 2001 (has links)
The sample dependency of item statistics and the practical importance of alternative models for test scores are evaluated in this case study using real data. The hypothesized superiority of the item response model (IRM) attributed to the sample invariance of item statistics is tested against a classical test theory (CTT) model and a structural equation model (SEM) for responses to the Center for Epidemiologic Studies-Depression (CES-D) scale. Sample-invariance of item statistics is tested in 10 random samples of 500 people, and across gender, age, and different health groups. Practical implications are considered in a comparison of differential score reliability, and individual rankings, using each test model. Item estimates from a 2-parameter logistic IRM were compared with classical item difficulty and discrimination estimates, and with item regression path estimates from a uni-factor SEM. An intraclass correlation coefficient (ICC) was calculated to evaluate the level of absolute agreement between item statistics in each condition. Item statistics from all test models were very similar across random samples, indicating a high level of invariance. However, IRM threshold parameters were least sensitive to sampling compared with other models. Greater variance was found among item statistics based on all test models across groups that differed in age, health, or gender. IRM discrimination estimates were most stable across contrasting groups, compared with those from other test models. Rankings assigned to individuals were most similar when CTT scores and linear transformed IRM scores were compared. The largest variation in individual rankings was obtained when SEM factor scores were compared with CTT scores in the higher score ranges. The reliability estimate for factor scores based on the SEM was highest overall. However, IRM optimal scores and the modified reliability estimate based on these, provide a more accurate estimate of average measurement error. This evidence supports the hypothesis of improved score precision when tests are constructed and scored using IRM techniques. However, rankings based on individual CES-D scores were very similar when the CTT, IRM, and SEM techniques were compared. Therefore, CTT or SEM scoring are reasonable alternatives to the IRM when norm-referenced score interpretations are based on CES-D scores.
|
100 |
Detecting DIF in polytomous items: An empirical comparison of the ordinal logistic regression, logistic discriminant function analysis, Mantel, and Generalized Mantel-Haenszel procedures.Kristjansson, Elizabeth. January 2001 (has links)
This study had four main objectives and two minor objectives: (1) To compare the Type I Error rates of four analytic techniques: the Mantel, Generalized Mantel-Haenszel (GMH), Logistic Discriminant Function Analysis (LDFA), and Ordinal Logistic Regression (OLR) procedures when there was no DIF in items. It was hypothesized that the procedures would differ little in their Type I Error rates, but that the OLR would have relatively the lowest Type I Error rates. (2) To compare the power of the Mantel, GMH, LDFA, and OLR for detecting uniform and nonuniform DIF in polytomous items. It was hypothesized that the Mantel would have the highest power for uniform DIF, but that it would not be useful for detecting nonuniform DIF. The GMH, OLR and LDFA were expected to display high power for nonuniform DIF. (3) To learn whether discrimination of the studied item, reference and focal group ability difference, sample size ratio between reference and focal group, and skewness would affect the performance of the four DIF detection procedures. It was hypothesized that differences in group ability distributions would result in increased Type I Error when item discrimination was high, and that differences in sample size ratio would lower power. (4) To determine whether adding a measure of effect size would reduce Type I Error. It was hypothesized that including effect size in the decision rule would reduce Type I Error for all procedures, and that it would also result in slightly lower power. (5) To compare Type I Error of the LDFA and OLR in classifying DIF as uniform when it was nonuniform and (6) To compare the Type I Error of the LDFA and OLR in classifying DIF as nonuniform when it was uniform. (Abstract shortened by UMI.)
|
Page generated in 0.1439 seconds