Spelling suggestions: "subject:"interraterreliability"" "subject:"interratereliability""
1 |
Interrater Agreement and Reliability of Observed Behaviors: Comparing Percentage Agreement, Kappa, Correlation Coefficient, ICC and G TheoryCao, Qian 02 October 2013 (has links)
The study of interrater agreement and itnerrater reliability attract extensive attention, due to the fact that the judgments from multiple raters are subjective and may vary individually. To evaluate interrater agreement and interrater reliability, five different methods or indices are proposed: percentage of agreement, kappa coefficient, the correlation coefficient, intraclass correlation coefficient (ICC), and generalizability (G) theory.
In this study, we introduce and discuss the advantages and disadvantages of these methods to evaluate interrater agreement and reliability. Then we review and explore the rank across these five indices by use of frequency in practice in the past five years. Finally, we illustrate how to use these five methods under different circumstances and provide SPSS and SAS code to analyze interrater agreement and reliability.
We apply the methods above to analyze the data from Parent-Child Interaction System of global ratings (PARCHISY), and conclude as follows: (1) ICC is the most often used method to evaluate interrater reliability in recent five years, while generalizability theory is the least often used method. The G coefficients provide similar interrater reliability with weighted kappa and ICC on most items, based on the criteria. (2) When the reliability is high itself, different methods provide consistent indication on interrater reliability based on different criteria. If the reliability is not consistent among different methods, both ICC and G coefficient will provide better interrater reliability based on the criteria, and they also provide consistent results.
|
2 |
Comparison of Video and Audio Rating Modalities for Assessment of Provider Fidelity to a Family-Centered, Evidence-Based ProgramJanuary 2019 (has links)
abstract: The current study assessed whether the interrater reliability and predictive validity of fidelity ratings differed significantly across the modalities of audio and video recordings. As empirically supported programs are moving to scale, attention to fidelity, the extent to which a program is delivered as intended, is essential because high fidelity is needed for positive program effects. Consequently, an important issue for prevention science is the development of feasible and acceptable methods for assessing fidelity. Currently, fidelity monitoring is rarely practiced, as the typical way of measuring fidelity, which uses video of sessions, is expensive, time-consuming, and intrusive. Audio recording has multiple advantages over video recording: 1) it is less intrusive; 2) equipment is less expensive; 3) recording procedures are simpler; 4) files are smaller so it takes less time to upload data and storage is less expensive; 5) recordings contain less identifying information; and 6) both clients and providers may be more willing to have sensitive interactions recorded with audio only. For these reasons, the use of audio recording may facilitate the monitoring of fidelity and increase the acceptability of both the intervention and implementation models, which may serve to broaden the scope of the families reached and improve the quality of the services provided. The current study compared the reliability and validity of fidelity ratings across audio and video rating modalities using 77 feedback sessions drawn from a larger randomized controlled trial of the Family Check-Up (FCU). Coders rated fidelity and caregiver in-session engagement at the age 2 feedback session. The composite fidelity and caregiver engagement scores were tested using path analysis to examine whether they predicted parenting behavior at age 3. Twenty percent of the sessions were double coded to assess interrater reliability. The interrater reliability and predictive validity of fidelity scores and caregiver engagement did not significantly differ across rating modality. However, caution must be used in interpreting these results because the interrater reliabilities in both conditions were low. Possible explanations for the low reliability, limitations of the current study, and directions for future research are discussed. / Dissertation/Thesis / Doctoral Dissertation Psychology 2019
|
3 |
Experts' Assessment of Color in Burn-Wound Photographs As a Predictor of Skin GraftBaker, Rose Ann Urdiales 01 July 2011 (has links)
No description available.
|
4 |
Assessment and Reporting of Intercoder Reliability in Published Meta-Analyses Related to Preschool Through Grade 12 EducationRaffle, Holly 10 October 2006 (has links)
No description available.
|
5 |
Interrater Reliability of the Psychological Rating Scale for Diagnostic ClassificationNicolette, Myrna 12 1900 (has links)
The poor reliability of the DSM diagnostic system has been a major issue of concern for many researchers and clinicians. Standardized interview techniques and rating scales have been shown to be effective in increasing interrater reliability in diagnosis and classification. This study hypothesized that the utilization of the Psychological Rating Scale for Diagnostic Classification for assessing the problematic behaviors, symptoms, or other characteristics of an individual would increase interrater reliability, subsequently leading to higher diagnostic agreement between raters and with DSM-III classification. This hypothesis was strongly supported by high overall profile reliability and individual profile reliability. Therefore utilization of this rating scale would enhance the accuracy of diagnosis and add to the educational efforts of technical personnel and those professionals in related disciplines.
|
6 |
Interrater reliabilita vyšetřovacího setu klinických funkcí u pacientů s roztroušenou sklerózou mozkomíšní / Interrater reliability of assessment set of clinical features of patients with multiple sclerosisMarková, Pavla January 2013 (has links)
Title: Interrater reliability of assessment set of clinical features of patients with multiple sclerosis Objectives: The aim of this thesis is to verify interrater reliability of the assessment set of clinical features of patients with multiple sclerosis which purpose is to evaluate sensitivly and comprehensivly the stage of the patients' clinical condition. Methods: According to the inclusion criteria, the patients with MS were selected by an independent neurologist who determined the EDSS score and duration of the disease. After, patients were evaluated by the assessment set by two independent physiotherapists. The assessment set of clinical features incluedes Low-Contrast Letter Acuity Test which tests contrast vision, Nine Hole Peg Test investiges fine motor skills, Timed 25 - Foot Walk evaluates walking speed over a distance of 7,5 m, Paced Auditory Serial Addition assesses cognitive function, Motricity Index tests muscle strength, Modified Ashworth Scale spasticity, Berg balance Scale equilibrium. Furthemore, the tests for evaluation of righting, equilibrium and protective reactions, the test evaluating knee hyperextension, the examination of dysdiadochokinesia and ataxia. Results: High interrater reliability was confirmed in all tests in the examinig set (ICC: 0,80 - 1), except for MAS...
|
7 |
Evaluierung bestehender Prüfungsmodalitäten in der Zahnärztlichen Vorprüfung und Implementierung neuer Prüfungsstrukturen / The evaluation of existing examination procedures of the dental preliminary exam and the implementation of a novel assessment toolEllerbrock, Maike 02 April 2019 (has links)
No description available.
|
8 |
Modified Profile Likelihood Approach for Certain Intraclass Correlation CoefficientLiu, Huayu 20 April 2011 (has links)
In this paper we consider the problem of constructing confidence intervals and lower bounds forthe intraclass correlation coefficient in an interrater reliability study where the raters are randomly selected from a population of raters.The likelihood function of the interrater reliability is derived and simplified, and the profile likelihood based approach is readily available for computing the confidence intervals of the interrater reliability. Unfortunately, the confidence intervals computed by using the profile likelihood function are in general too narrow to have the desired coverage probabilities. From the point view of practice, a conservative approach, if is at least as precise as any existing method, is preferred sinceit gives the correct results with a probability higher than claimed. Under this rationale, we propose the so-called modified likelihood approach in this paper. Simulation study shows that, the proposed method in general has better performance than currently used methods.
|
9 |
Interrater Reliability Of Psychomotor Skill Assessment In Athletic TrainingCraddock, Jason 01 January 2009 (has links)
Assessment in athletic training education is an evolutionary process that is determined by each individual Athletic Training Education Program. The autonomy authorized by national accreditation standards allows academic programs to determine the appropriate assessment practices that facilitate the meeting of student learning outcomes. Even with autonomy, formative and summative techniques are to be employed in both the didactic and clinical arenas of athletic training education programs. The major objective of athletic training education is to prepare students for entry-level practice in athletic training. The purpose of this study was to assess interrater reliability of athletic training faculty and approved clinical instructors in their rating of athletic training student performance on four psychomotor skills. A total of 115 individuals participated in this study. Thirty two faculty and 83 approved clinical instructors completed the online survey The results of this study indicate that the overall reliability was high for the entire population as well as the subgroups analyzed. Even though the overall reliability was high, three specific criteria out of a total of 29 criteria had lower reliability scores. These findings may indicate that there may be a high degree of agreement between academic faculty and approved clinical instructors in the rating of athletic training students in their performance of psychomotor skills.
|
10 |
Contributions to estimation of measures for assessing rater reliabilityWang, Luqiang January 2009 (has links)
Reliability measures have been well studied over many years, beginning with an entire chapter devoted to intraclass correlation in the first edition of Fisher (1925). Such measures have been thoroughly studied for two factor models. This dissertation, motivated by a medical research problem, extends point and confidence interval estimation of both intraclass correlation coefficient and interater reliability coefficient to models containing three crossed random factors -- subjects, raters and occasions. The intraclass correlation coefficient is used when decision is made on an absolute basis with rater's scores, while the interater reliability coefficient is defined for decisions made on a relative basis. The estimation is conducted using both ANOVA and MCMC methods. The results from the two methods are compared. The MCMC method is preferred for analyses of small data sets when ICC values are high. Besides, the bias of estimator of intraclass correlation coefficient in one-way random effects model is evaluated. / Statistics
|
Page generated in 0.0972 seconds