首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
The central purposes of this study were to review the development and evolution of the Scientific Attitude Inventory (SAI) and then reevaluate the psychometric properties of the revised form of the SAI, the Scientific Attitude Inventory II (SAI‐II). The SAI‐II was administered to a convenience sample of 543 middle and high school students from five teachers in four schools in four school districts in San Antonio, Texas, at the beginning of the 2004–2005 school year. Confirmatory factor analysis on the full data set failed to support the existence of a 12‐factor structure (as proposed by the scale developers) or a one‐factor structure. The data were then randomly divided into exploratory [exploratory factor analysis (EFA)] validation and confirmatory [confirmatory factor analysis (CFA)] cross‐validation sets. Exploratory and confirmatory models yielded a three‐factor solution that did not fit the data well [χ2 (321) = 646, p < .001; RMSEA = .061 (.90 CI = .054–.068); and CFI = .81]. The three factors were labeled “Science is About Understanding and Explaining” (13 items), “Science is Rigid” (6 items), and “I Want to Be a Scientist” (8 items). The α‐coefficients for these three factors ranged from 0.59 to 0.85. Whether these identified subscales are valid will require independent investigation. In this sample, and consistent with prior publications, the SAI‐II in its current form did not have satisfactory psychometric properties and cannot be recommended for further use. © 2008 Wiley Periodicals, Inc. J Res Sci Teach 45: 600–616, 2008  相似文献   

2.
Because the psychological assessment of high ability usually concentrates on intelligence testing, it is pertinent to discuss the validity of intelligence test batteries. The well‐known Wechsler's scales are analyzed and evaluated. Based on psychometric models, especially the Rasch model, analyses are made of some German editions, which show that hardly a single subtest scores fairly. That is, the true extent of testees’ abilities will not be correctly represented by the scores obtained under current scoring rules. Since many of the items of the analyzed editions correspond to items of the American edition (WISC‐R), the same shortcomings must also be suspect for that test battery. In this light, the administration of these tests is no longer acceptable. However, it is shown that Wechsler's basic concept is worthwhile when accompanied by (modern) psychometric tools: a new (German) test battery, AID, is introduced which, in particular, conforms to economic requirements if high ability is to be assessed.  相似文献   

3.
4.
The Peabody Individual Achievement Test–Revised (PIAT‐R) and Wechsler Individual Achievement Test (WIAT) contain measures of written expression. However, these subtests are not theory‐based and were assessed with inappropriate psychometric analyses. This study attempted to enhance the study of written expression by reexamining the reliability and validity of the PIAT‐R and WIAT Written Expression scoring systems, applying theory and more appropriate statistical analyses. First, items were identified that were the most and least reliable, determined by interrater agreement. Next, the most and least valid items were identified, based on item–total correlations. Subjects included 50 adolescents, men, and women aged 13 to 46 years; raters were three graduate students with experience and training similar to that of the typical test user. Results indicate that seven items were too easy, as virtually all subjects received the maximum score on these items—these items were eliminated. The remaining 24 items were classified as both reliable and valid (9 items), reliable but not valid (4 items), valid with limited reliability (5 items), and neither reliable nor valid (6 items). The WIAT written expression scoring system was found to have more items that were both reliable and valid compared to the PIAT‐R scoring system. Items measuring global, rather than specific, content were also found to be more reliable and valid. © 1999 John Wiley & Sons, Inc.  相似文献   

5.
In this article the authors present a confirmatory factor analysis of the Teachers’ Attitudes Toward Computers (TAC) and the Teachers’ Attitudes Toward Information Technology (TAT) scales by Christensen and Knezek (1996, 1998) using large samples from three states. The TAC was reduced from 98 items and nine factors to 35 items and eight factors, while maintaining superior psychometric properties. The TAT was reduced from 50 items and five factors to 20 items measuring five factors, similarly maintaining excellent psychometric properties. The authors also developed a combined short form of both instruments using 42 items to measure 11 factors for use in situations where parsimony is critical.  相似文献   

6.
The purpose of this study is to apply the attribute hierarchy method (AHM) to a subset of SAT critical reading items and illustrate how the method can be used to promote cognitive diagnostic inferences. The AHM is a psychometric procedure for classifying examinees’ test item responses into a set of attribute mastery patterns associated with different components from a cognitive model. The study was conducted in two steps. In step 1, three cognitive models were developed by reviewing selected literature in reading comprehension as well as research related to SAT Critical Reading. Then, the cognitive models were validated by having a sample of students think aloud as they solved each item. In step 2, psychometric analyses were conducted on the SAT critical reading cognitive models by evaluating the model‐data fit between the expected and observed response patterns produced from two random samples of 2,000 examinees who wrote the items. The model that provided best data‐model fit was then used to calculate attribute probabilities for 15 examinees to illustrate our diagnostic testing procedure.  相似文献   

7.
The aim of this article is twofold: (a) to validate the psychometric properties of the Academic Resilience Scale (AR‐S) in a Spanish university context; and (b) to test a model where different coping strategies are antecedents of AR, and where academic satisfaction and performance are its consequences. The studies were conducted with 185 (study 1) and 780 (study 2) students. In study 1, the construct validity, reliability, and convergent and divergent validity of the AR‐S were analyzed. The hypothesized model (study 2) was tested using structural equation modeling. The results confirmed the good psychometric properties of the AR‐S in the Spanish context, providing evidence of its validity. Furthermore, the distinctive role of different coping strategies in resilience was confirmed, as well as the relationship between resilience and academic satisfaction. No direct relationship between resilience and performance was found, highlighting the full mediation of satisfaction.  相似文献   

8.
Approximately 30% of students drop out from Norwegian upper secondary schools. Academic achievement, as indexed by grade point average (GPA), is one of the strongest predictors of dropout. The present study aimed to examine the role of cognitive, school-related and affective/psychological predictors of GPA. In addition, we examined the psychometric properties of a new scale for literacy problems (LP), as well as its role for GPA. Analyses showed that the scale had favourable psychometric properties. Multivariable regression analyses showed that study track choice, task solving skills, gender, lack of educational plans and LP predicted GPA in descending order. The major implication of the study was that GPA is related to a multiple number of demographic, cognitive and psychosocial factors, and that any interventions addressing GPA will be less effective if not addressing psychosocial in addition to cognitive/school-related factors.  相似文献   

9.
OBJECTIVE: The goal was to develop a retrospective inventory of parental threatening behavior to facilitate a better understanding of such behavior's role in the etiology of psychological distress. METHOD: Inventory items were developed based on theory and 135 students' responses to a question eliciting examples of threatening parental behavior. Following item development, two additional student samples (n = 200 and n = 603) completed batteries of self-report measures. Responses were used to eliminate unstable or redundant items from the inventory and to examine the inventory's psychometric properties. RESULTS: Factor analysis of the inventory revealed three factors, accounting for 66.2% of variance; this factor structure is compatible with theory, and consistent across maternal behavior scores, paternal behavior scores, and combined maternal and paternal scores. Cronbach's coefficient alphas indicated acceptable internal consistency; Pearson correlation coefficients indicated acceptable 4-week test-retest reliability. Moderate intercorrelations with two retrospective measures of childhood experiences suggested construct validity. Regression analyses demonstrated the ability of the inventory to predict both anxious and depressive symptomatology and lifetime symptoms of anxiety and depressive disorder. Normative data on combined parent scores, maternal scores, and paternal scores are also presented. CONCLUSIONS: Initial psychometric testing of the Parent Threat Inventory (PTI) suggests it is a reliable and valid tool for investigating the developmental antecedents of adult psychological distress. Further research should focus on addressing two limitations: (1) lack of normative and psychometric data on men and women suffering from clinical disorders, and (2) lack of validation by parental reporting.  相似文献   

10.
The Arnett Caregiver Interaction Scale (CIS) has been widely used in research studies to measure the quality of caregiver–child interactions. The scale was modeled on a well-established theory of parenting, but there are few psychometric studies of its validity. We applied factor analyses and item response theory methods to assess the psychometric properties of the Arnett CIS in a national sample of toddlers in home-based care and preschoolers in center-based care from the Early Childhood Longitudinal Study-Birth Cohort. We found that a bifactor structure (one common factor and a second set of specific factors) best fits the data. In the Arnett CIS, the bifactor model distinguishes a common substantive dimension from two methodological dimensions (for positively and negatively oriented items). Despite the good fit of this model, the items are skewed (most teachers/caregivers display positive interactions with children) and, as a result, the Arnett CIS is not well suited to distinguish between caregivers who are “highly” versus “moderately” positive in their interactions with children, according to the items on the scale. Regression-adjusted associations between the Arnett CIS and child outcomes are small, especially for preschoolers in centers. We encourage future scale development work on measures of child care quality by early childhood scholars.  相似文献   

11.
12.
Using Rasch analysis, the psychometric properties of a newly developed 35‐item parent‐proxy instrument, the Caregiver Assessment of Movement Participation (CAMP), designed to measure movement participation problems in children with Developmental Coordination Disorder, were examined. The CAMP was administered to 465 school children aged 5–10 years. Thirty of the 35 items were retained as they had acceptable infit and outfit statistics. Item separation (7.48) and child separation (3.16) were good; moreover, the CAMP had excellent reliability (Reliability Index for item = 0.98; Person = 0.91). Principal components analysis of item residuals confirmed the unidimensionality of the instrument. Based on category probability statistics, the original five‐point scale was collapsed into a four‐point scale. The item threshold calibration of the CAMP with the Movement Assessment Battery for Children Test was computed. The results indicated that a CAMP total score of 75 is the optimal cut‐off point for identifying children at risk of movement problems.  相似文献   

13.
《Educational Assessment》2013,18(3):201-224
This article discusses an approach to analyzing performance assessments that identifies potential reasons for misfitting items and uses this information to improve on items and rubrics for these assessments. Specifically, the approach involves identifying psychometric features and qualitative features of items and rubrics that may possibly influence misfit; examining relations between these features and the fit statistic; conducting an analysis of student responses to a sample of misfitting items; and finally, based on the results of the previous analyses, modifying characteristics of the items or rubrics and reexamining fit. A mathematics performance assessment containing 53 constructed-response items scored on a holistic scale from 0 to 4 is used to illustrate the approach. The 2-parameter graded response model (Samejima, 1969) is used to calibrate the data. Implications of this method of data analysis for improving performance assessment items and rubrics are discussed as well as issues and limitations related to the use of the approach.  相似文献   

14.
Angus Duff 《教育心理学》2004,24(5):699-709
Given the psychometric limitations of existing measures of Kolb's experiential learning model (ELM), two new scales of learning styles have been developed. The validity of these scales has been supported in samples of undergraduate and MBA students in the USA. This paper provides evidence of some psychometric properties of scores yielded by these scales using samples of undergraduate students in the UK. Only limited support is found for the internal consistency reliability and construct validity of scores produced by the scales. However, an item attrition exercise identifies a two‐factor solution providing an acceptable fit to the data. The scales are reported as being positively correlated with academic performance and prior academic achievement. Despite the mixed evidence, we suggest further development of the scales is warranted to create a psychometrically sound measure of the ELM.  相似文献   

15.
The continual supply of new items is crucial to maintaining quality for many tests. Automatic item generation (AIG) has the potential to rapidly increase the number of items that are available. However, the efficiency of AIG will be mitigated if the generated items must be submitted to traditional, time‐consuming review processes. In two studies, generated mathematics achievement items were subjected to multiple stages of qualitative review for measuring the intended skills, followed by empirical tryout in operational testing. High rates of success were found. Further, items generated from the same item structure had predictable psychometric properties. Thus, the feasibility of a more limited and expedient review processes was supported. Additionally, positive results were obtained on measuring the same skills from item structures with reduced cognitive complexity.  相似文献   

16.
The psychometric literature provides little empirical evaluation of examinee test data to assess essential psychometric properties of innovative items. In this study, examinee responses to conventional (e.g., multiple choice) and innovative item formats in a computer-based testing program were analyzed for IRT information with the three-parameter and graded response models. The innovative item types considered in this study provided more information across all levels of ability than multiple-choice items. In addition, accurate timing data captured via computer administration were analyzed to consider the relative efficiency of the multiple choice and innovative item types. As with previous research, multiple-choice items provide more information per unit time. Implications for balancing policy, psychometric, and pragmatic factors in selecting item formats are also discussed.  相似文献   

17.
The study investigated the psychometric properties of the Eysenck Personality Questionnaire – Revised (EPQ‐R) Short Scale administered to two groups of Brunei trainee teachers of both genders taking psychology courses in English (117 students) and Malay (106 students). The extraversion scale had the most problematic items. Both administrations produced means and standard deviations comparable to the norms in the technical manual. All the scales, except P (psychoticism), had satisfactory levels of alpha reliability and validity (discriminant and construct) on both versions. In addition factor analysis correctly assigned some of the items to their respective scales on both versions. No statistically significant gender difference in performance on both scales was obtained. Findings suggest that the instrument was suitable for use with Brunei student teachers with tertiary‐level education when technical terms were clarified and reasonable administration time was given.  相似文献   

18.
The psychometric properties of a 10-item math motivation scale were empirically validated with an independent sample consisting of 182 elementary-school students. Analysis of the model dimensionality supported a one-factor structure fit. Item parameter estimates from a Classical Test Theory framework revealed that most items were highly discriminating, and the survey is informative for students of low to average math motivation. Differential item functioning (DIF) analyses found two items exhibiting gender bias. Overall, the instrument was a psychometrically valid instrument for measuring math motivation at the elementary level.  相似文献   

19.
It is of vital importance that children are educated in a safe environment. Every school needs to have a well‐developed crisis management document containing plans for prevention, intervention, and postvention. We developed the Comprehensive Crisis Plan Checklist (CCPC) to serve as a valuable tool that can be used to assist practitioners with evaluation of their school crisis management plans. To use the CCPC, school personnel should match the items on the checklist to elements in their developing or existing crisis plan to determine missing elements. By comparing a school or district crisis plan to the CCPC, practitioners or school personnel will be able to determine strengths and weaknesses of the plan as well as provide feedback for making their plans more complete and comprehensive. In this study, we describe the development of the CCPC and report on reliability data using this checklist. © 2010 Wiley Periodicals, Inc.  相似文献   

20.
This paper illustrates that the psychometric properties of scores and scales that are used with mixed‐format educational tests can impact the use and interpretation of the scores that are reported to examinees. Psychometric properties that include reliability and conditional standard errors of measurement are considered in this paper. The focus is on mixed‐format tests in situations for which raw scores are integer‐weighted sums of item scores. Four associated real‐data examples include (a) effects of weights associated with each item type on reliability, (b) comparison of psychometric properties of different scale scores, (c) evaluation of the equity property of equating, and (d) comparison of the use of unidimensional and multidimensional procedures for evaluating psychometric properties. Throughout the paper, and especially in the conclusion section, the examples are related to issues associated with test interpretation and test use.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号